Principles of
Tensor Calculus
gi j = Ei · E j
∂r
∂ ui
u1 CS
∇2
f=
δ Ak
δt
√1
g ∂i
=
E2
j
E1
i
E1
u2 CC
Aj
E2
=A
i
E3
u2 CS
j1 ... jn
u1 CC
√
gg i j
∂j f
dAk
dt
δ jin1
δ jin2
..
.
δ jinn
···
···
..
.
···
E3
...in
δ ji11 ...
jn =
ε i1 ...in ε
δ ji21
δ ji22
..
.
δ ji2n
δ
Ei =
...in
δ ji11 ...
jn =
u3 CC
δ ji11
δ ji12
..
.
δ ji1n
du j
i
k
+ Γi j A dt
u3 CS
i
j
g
i
E
i
∇u
=
=
i
j
E ·E
Taha Sochi
Preface
This book is based on my previous book: Tensor Calculus Made Simple, where the development of tensor calculus concepts and techniques are continued at a higher level. In
the present book, we continue the discussion of the main topics of the subject at a more
advanced level expanding, when necessary, some topics and developing further concepts
and techniques. The purpose of the present book is to solidify, generalize, fill the gaps and
make more rigorous what have been presented in the previous book.
Unlike the previous book which is largely based on a Cartesian approach, the formulation
in the present book is largely based on assuming an underlying general coordinate system
although some example sections are still based on a Cartesian approach for the sake of
simplicity and clarity. The reader will be notified about the underlying system in the given
formulation. We also provide a sample of formal proofs to familiarize the reader with the
tensor techniques. However, due to the preset objectives and the intended size of the book,
we do not offer comprehensive proofs and complete theoretical foundations for the provided
materials although we generally try to justify many of the given formulations descriptively
or by interlinking to related formulations or by similar pedagogical techniques. This may
be seen as a more friendly method for constructing and establishing the abstract concepts
and techniques of tensor calculus.
The book is furnished with an index in the end of the book as well as rather detailed sets
of exercises in the end of each chapter to provide useful revision and practice. To facilitate
linking related concepts and sections, and hence ensure better understanding of the given
materials, cross referencing, which is hyperlinked for the ebook users, is used extensively
throughout the book. The book also contains a number of graphic illustrations to help
the readers to visualize the ideas and understand the subtle concepts.
The book can be used as a text for an introductory or an intermediate level course on
tensor calculus. The familiarity with the materials presented in the previous book will be
an advantage although it is not necessary for someone with a reasonable mathematical
background. Moreover, the main materials of the previous book are absorbed within the
structure of the present book for the sake of completeness and to make the book rather
self-contained considering the predetermined objectives. I hope I achieved these goals.
Taha Sochi
London, August 2017
1
Contents
Preface
1
Table of Contents
2
Nomenclature
6
1 Preliminaries
1.1 General Conventions and Notations . . . . . . . . . . . . . . . . . . . . . .
1.2 General Background about Tensors . . . . . . . . . . . . . . . . . . . . . .
1.3 Exercises and Revision . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9
9
12
19
2 Spaces, Coordinate Systems and Transformations
2.1 Spaces . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.2 Coordinate Systems . . . . . . . . . . . . . . . . . . . . . . .
2.2.1 Rectilinear and Curvilinear Coordinate Systems . . .
2.2.2 Orthogonal Coordinate Systems . . . . . . . . . . . .
2.2.3 Homogeneous Coordinate Systems . . . . . . . . . . .
2.3 Transformations . . . . . . . . . . . . . . . . . . . . . . . . .
2.3.1 Proper and Improper Transformations . . . . . . . .
2.3.2 Active and Passive Transformations . . . . . . . . . .
2.3.3 Orthogonal Transformations . . . . . . . . . . . . . .
2.3.4 Linear and Nonlinear Transformations . . . . . . . .
2.4 Coordinate Curves and Coordinate Surfaces . . . . . . . . .
2.5 Scale Factors . . . . . . . . . . . . . . . . . . . . . . . . . .
2.6 Basis Vectors and Their Relation to the Metric and Jacobian
2.7 Relationship between Space, Coordinates and Metric . . . .
2.8 Exercises and Revision . . . . . . . . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
22
22
24
25
26
27
29
32
33
33
34
35
36
39
44
46
3 Tensors
3.1 Tensor Types . . . . . . . . . . . . . . . . . . .
3.1.1 Covariant and Contravariant Tensors . .
3.1.2 True and Pseudo Tensors . . . . . . . . .
3.1.3 Absolute and Relative Tensors . . . . . .
3.1.4 Isotropic and Anisotropic Tensors . . . .
3.1.5 Symmetric and Anti-symmetric Tensors .
3.1.6 General and Affine Tensors . . . . . . . .
3.2 Tensor Operations . . . . . . . . . . . . . . . .
3.2.1 Addition and Subtraction . . . . . . . .
3.2.2 Multiplication of Tensor by Scalar . . . .
3.2.3 Tensor Multiplication . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
50
50
50
55
57
58
59
62
62
63
64
64
2
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
66
66
68
69
69
71
4 Special Tensors
4.1 Kronecker delta Tensor . . . . . . . . . . . . . . . . . . . . . . .
4.2 Permutation Tensor . . . . . . . . . . . . . . . . . . . . . . . . .
4.3 Identities Involving Kronecker or/and Permutation Tensors . . .
4.3.1 Identities Involving Kronecker delta Tensor . . . . . . . .
4.3.2 Identities Involving Permutation Tensor . . . . . . . . .
4.3.3 Identities Involving Kronecker and Permutation Tensors
4.4 Generalized Kronecker delta Tensor . . . . . . . . . . . . . . . .
4.5 Metric Tensor . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4.6 Definitions Involving Special Tensors . . . . . . . . . . . . . . .
4.6.1 Dot Product . . . . . . . . . . . . . . . . . . . . . . . . .
4.6.2 Magnitude of Vector . . . . . . . . . . . . . . . . . . . .
4.6.3 Angle between Vectors . . . . . . . . . . . . . . . . . . .
4.6.4 Cross Product . . . . . . . . . . . . . . . . . . . . . . . .
4.6.5 Scalar Triple Product . . . . . . . . . . . . . . . . . . . .
4.6.6 Vector Triple Product . . . . . . . . . . . . . . . . . . .
4.6.7 Determinant of Matrix . . . . . . . . . . . . . . . . . . .
4.6.8 Length . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4.6.9 Area . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4.6.10 Volume . . . . . . . . . . . . . . . . . . . . . . . . . . .
4.7 Exercises and Revision . . . . . . . . . . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
76
76
77
82
82
83
85
88
91
97
97
98
98
99
100
101
102
102
103
105
107
5 Tensor Differentiation
5.1 Christoffel Symbols . . .
5.2 Covariant Differentiation
5.3 Absolute Differentiation
5.4 Exercises and Revision .
3.3
3.4
3.2.4 Contraction . .
3.2.5 Inner Product .
3.2.6 Permutation . .
3.2.7 Tensor Test and
Tensor Representations
Exercises and Revision
. . . . . . . . .
. . . . . . . . .
. . . . . . . . .
Quotient Rule
. . . . . . . . .
. . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
112
112
120
129
133
6 Differential Operations
6.1 Cartesian Coordinate System
6.1.1 Operators . . . . . . .
6.1.2 Gradient . . . . . . . .
6.1.3 Divergence . . . . . . .
6.1.4 Curl . . . . . . . . . .
6.1.5 Laplacian . . . . . . .
6.2 General Coordinate System .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
137
138
138
138
139
139
140
140
.
.
.
.
.
.
.
.
3
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
140
140
142
143
144
146
146
147
147
147
148
148
148
149
149
149
149
150
150
150
151
151
151
152
7 Tensors in Application
7.1 Tensors in Mathematics . . . . . . . . . . . . . . . . . . . . . .
7.1.1 Common Definitions in Tensor Notation . . . . . . . . .
7.1.2 Scalar Invariants of Tensors . . . . . . . . . . . . . . . .
7.1.3 Common Identities in Vector and Tensor Notation . . . .
7.1.4 Integral Theorems in Tensor Notation . . . . . . . . . . .
7.1.5 Examples of Using Tensor Techniques to Prove Identities
7.2 Tensors in Geometry . . . . . . . . . . . . . . . . . . . . . . . .
7.2.1 Riemann-Christoffel Curvature Tensor . . . . . . . . . .
7.2.2 Bianchi Identities . . . . . . . . . . . . . . . . . . . . . .
7.2.3 Ricci Curvature Tensor and Scalar . . . . . . . . . . . .
7.3 Tensors in Science . . . . . . . . . . . . . . . . . . . . . . . . . .
7.3.1 Infinitesimal Strain Tensor . . . . . . . . . . . . . . . . .
7.3.2 Stress Tensor . . . . . . . . . . . . . . . . . . . . . . . .
7.3.3 Displacement Gradient Tensors . . . . . . . . . . . . . .
7.3.4 Finger Strain Tensor . . . . . . . . . . . . . . . . . . . .
7.3.5 Cauchy Strain Tensor . . . . . . . . . . . . . . . . . . . .
7.3.6 Velocity Gradient Tensor . . . . . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
155
155
156
157
158
160
161
167
167
171
172
173
173
174
174
175
175
175
6.3
6.4
6.5
6.6
6.2.1 Operators . . . . . . . .
6.2.2 Gradient . . . . . . . . .
6.2.3 Divergence . . . . . . . .
6.2.4 Curl . . . . . . . . . . .
6.2.5 Laplacian . . . . . . . .
Orthogonal Coordinate System
6.3.1 Operators . . . . . . . .
6.3.2 Gradient . . . . . . . . .
6.3.3 Divergence . . . . . . .
6.3.4 Curl . . . . . . . . . . .
6.3.5 Laplacian . . . . . . . .
Cylindrical Coordinate System .
6.4.1 Operators . . . . . . . .
6.4.2 Gradient . . . . . . . . .
6.4.3 Divergence . . . . . . . .
6.4.4 Curl . . . . . . . . . . .
6.4.5 Laplacian . . . . . . . .
Spherical Coordinate System . .
6.5.1 Operators . . . . . . . .
6.5.2 Gradient . . . . . . . . .
6.5.3 Divergence . . . . . . . .
6.5.4 Curl . . . . . . . . . . .
6.5.5 Laplacian . . . . . . . .
Exercises and Revision . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
4
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
7.4
7.3.7 Rate of Strain Tensor . . . . . . . . . . . . . . . . . . . . . . . . . . 176
7.3.8 Vorticity Tensor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 176
Exercises and Revision . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 177
References
180
Index
181
Author Notes
187
5
Nomenclature
In the following list, we define the common symbols, notations and abbreviations which
are used in the book as a quick reference for the reader.
∇
∇; and ∇;
∇f
∇·A
∇×A
∇2 , ∂ii , ∇ii
∇v, ∂i vj
, (subscript)
; (subscript)
hat (e.g. Âi , Êi )
bar (e.g. ūi , Āi )
◦
⊥
1D, 2D, 3D, nD
δ/δt
∂i and ∇i
∂;i
[ij, k]
A
B, Bij
B−1 , Bij−1
C
Cn
d, di
det
dr
ds
dσ
dτ
ei
er , eθ , eφ
err , erθ , · · · , eφφ
eρ , eφ , ez
eρρ , eρφ , · · · , ezz
E, Eij
Ei , Ei
Ei
nabla differential operator
covariant and contravariant differential operators
gradient of scalar f
divergence of tensor A
curl of tensor A
Laplacian operator
velocity gradient tensor
partial derivative with respect to following index(es)
covariant derivative with respect to following index(es)
physical representation or normalized vector
transformed quantity
inner or outer product operator
perpendicular to
one-, two-, three-, n-dimensional
absolute derivative operator with respect to t
partial derivative operator with respect to ith variable
covariant derivative operator with respect to ith variable
Christoffel symbol of 1st kind
area
Finger strain tensor
Cauchy strain tensor
curve
of class n
displacement vector
determinant of matrix
differential of position vector
length of infinitesimal element of curve
area of infinitesimal element of surface
volume of infinitesimal element of space
ith vector of orthonormal vector set (usually Cartesian basis set)
basis vectors of spherical coordinate system
unit dyads of spherical coordinate system
basis vectors of cylindrical coordinate system
unit dyads of cylindrical coordinate system
first displacement gradient tensor
ith covariant and contravariant basis vectors
ith orthonormalized covariant basis vector
6
Eq./Eqs.
g
g
gij , g ij , gij
g11 , g12 , · · · gnn
g 11 , g 12 , · · · g nn
hi
iff
J
J
J−1
L
n, ni
P
P (n, k)
qi
qi
r
R
Rij , Rij
Rijkl , Rijkl
r, θ, φ
S
S, Sij
S̄, S̄ij
t
T (superscript)
T, Ti
tr
ui
v, vi
V
w
xi , xi
′
xi , xi
x, y, z
γ, γij
γ̇
Γkij
δ
δij , δ ij , δij
ijk
ij
n
, δji11 ...i
, δlmn
δkl
...jn
∆, ∆ij
Equation/Equations
determinant of covariant metric tensor
metric tensor
covariant, contravariant and mixed metric tensor or its components
coefficients of covariant metric tensor
coefficients of contravariant metric tensor
scale factor for ith coordinate
if and only if
Jacobian of transformation between two coordinate systems
Jacobian matrix of transformation between two coordinate systems
inverse Jacobian matrix of transformation
length of curve
normal vector to surface
point
k-permutations of n objects
ith coordinate of orthogonal coordinate system
ith unit basis vector of orthogonal coordinate system
position vector
Ricci curvature scalar
Ricci curvature tensor of 1st and 2nd kind
Riemann-Christoffel curvature tensor of 1st and 2nd kind
coordinates of spherical coordinate system
surface
rate of strain tensor
vorticity tensor
time
transposition of matrix
traction vector
trace of matrix
ith coordinate of general coordinate system
velocity vector
volume
weight of relative tensor
ith Cartesian coordinate
ith Cartesian coordinate of particle at past and present times
coordinates of 3D space (mainly Cartesian)
infinitesimal strain tensor
rate of strain tensor
Christoffel symbol of 2nd kind
Kronecker delta tensor
covariant, contravariant and mixed ordinary Kronecker delta
generalized Kronecker delta in 2D, 3D and nD space
second displacement gradient tensor
7
ǫij , ǫijk , ǫi1 ...in
ǫij , ǫijk , ǫi1 ...in
ǫij , ǫijk , ǫi1 ...in
ǫij , ǫijk ,ǫi1 ...in
ρ, φ
ρ, φ, z
σ, σij
ω
Ω
covariant relative permutation tensor in 2D, 3D and nD space
contravariant relative permutation tensor in 2D, 3D and nD space
covariant absolute permutation tensor in 2D, 3D and nD space
contravariant absolute permutation tensor in 2D, 3D and nD space
coordinates of plane polar coordinate system
coordinates of cylindrical coordinate system
stress tensor
vorticity tensor
region of space
8
Chapter 1
Preliminaries
In this introductory chapter, we provide preliminary materials about conventions and
notations as well as basic facts about tensors which will be needed in the subsequent parts
of the book. The chapter is therefore divided into two sections: the first is about general
conventions and notations used in the book, and the second is on general background
about tensors.
1.1
General Conventions and Notations
In this section, we provide general notes about the main conventions and notations used in
the present book. We usually use the term “tensor” to mean tensors of all ranks including
scalars (rank-0) and vectors (rank-1). However, we may also use this term as opposite to
scalar and vector, i.e. tensor of rank-n where n > 1. In almost all cases, the meaning
should be obvious from the context. We note that in the present book all tensors of all
ranks and types are assumed to be real quantities, i.e. they have real rather than imaginary
or complex components.
We use non-indexed lower case light face italic Latin letters (e.g. f and h) to label
scalars, while we use non-indexed lower or upper case bold face non-italic Latin letters
(e.g. a and A) to label vectors in symbolic notation. The exception to this is the basis
vectors where indexed bold face lower or upper case non-italic symbols (e.g. e1 and Ei ) are
used. However, there should be no confusion or ambiguity about the meaning of any one
of these symbols. We also use non-indexed upper case bold face non-italic Latin letters
(e.g. A and B) to label tensors of rank > 1 in symbolic notation. Since matrices in
this book are supposed to represent rank-2 tensors, they also follow the rules of labeling
tensors symbolically by using non-indexed upper case bold face non-italic Latin letters.
We note that in a few cases in the final chapter (see § 7.3) we used boldface and indexed
light face Greek symbols to represent particular tensors, which are commonly labeled in
the literature by these symbols, to keep with the tradition.
Indexed light face italic Latin symbols (e.g. ai and Bijk ) are used in this book to denote
tensors of rank > 0 in their explicit tensor form, i.e. index notation. Such symbols may
also be used to denote the components of these tensors. The meaning is usually transparent
and can be identified from the context if it is not declared explicitly. Tensor indices in this
book are lower case Latin letters which may be taken preferably from the middle of the
Latin alphabet (such as i, j and k) for the free indices and from the beginning of the Latin
alphabet (such as a and b) for the dummy indices. We also use numbered indices, such
as (i1 , i2 , . . . , ik ), for this purpose when the number of tensor indices is variable. Numbers
are also used as indices in some occasions (e.g. ǫ12 ) for obvious purposes such as making
statements about particular components.
9
1.1 General Conventions and Notations
10
Partial derivative symbol with a subscript index (e.g. ∂i ) is used to denote partial
differentiation with respect to the ith variable, that is:
∂i =
∂
∂xi
(1)
However,we should note that in this book we generalize partial derivative notation so that
∂i symbolizes partial derivative with respect to the ui coordinate of general coordinate
systems and not just Cartesian coordinates which are usually denoted by xi or xi . The
type of coordinates, being Cartesian or general or otherwise, will be determined by the
context which should be obvious in all cases.
Similarly, we use partial derivative symbol with a twice-repeated index to denote the
Laplacian operator, that is:[1]
∂ii = ∂i ∂i = ∇2
(2)
Partial derivative symbol with a coordinate label subscript, rather than an index, is also
used to denote partial differentiation with respect to that spatial variable. For instance:
∂r =
∂
∂r
(3)
is used to denote the partial derivative with respect to the radial coordinate r in spherical
coordinate systems which are identified by the spatial variables (r, θ, φ). It should be
obvious that in notations like ∂r the subscript is used as a label rather than an index and
hence it does not follow the rules of tensor indices which will be discussed later (see § 1.2).
Following the widely used convention, a subscript comma preceding a subscript index
(e.g. Ak,i ) is used to denote partial differentiation with respect to the spatial coordinate
which is indexed by the symbol that follows the comma. For example, f,i and Ajk,i are
used to represent the partial derivative of the scalar f and rank-2 tensor Ajk with respect
to the ith coordinate, that is:
f,i = ∂i f
Ajk,i = ∂i Ajk
(4)
We also follow the common convention of using a subscript semicolon preceding a subscript
index (e.g. Akl;i ) to symbolize the operation of covariant differentiation with respect to
the ith coordinate (see § 5.2). The semicolon notation may also be attached to the normal
differential operators for the same purpose. For example, ∇;i and ∂;i symbolize covariant
differential operators with respect to the ith variable.
In this regard, we should remark that more than one index may follow the comma and
semicolon in these notations to represent multiple partial and covariant differentiation with
respect to the indexed variables according to the stated order of the indices. For example,
Ai,jk is used to represent the mixed second order partial derivative of the tensor Ai with
respect to the j th and k th coordinates, while Bji;km is used to represent the mixed second
[1]
This is the Cartesian form. For the other forms, the reader is referred to § 6.
1.1 General Conventions and Notations
11
order covariant derivative of the tensor Bji with respect to the k th and mth coordinates,
that is:
Bji;km = ∇;m (∇;k Bji )
(5)
Ai,jk = ∂k ∂j Ai
We also note that in a few occasions superscripts, rather than subscripts, comma and semicolon preceding superscript index (e.g. f ,i and Ai;j ) are used to represent contravariant
partial derivative and contravariant tensor derivative respectively. Superscripted differential operators (e.g. ∂ i and ∇;i ) are also used occasionally to represent these differential
operators in their contravariant form. A matter related to tensor differentiation is that
we follow the conventional notation δtδ to represent the intrinsic derivative, which is also
known as the absolute derivative, with respect to the variable t along a given curve, as
will be discussed in § 5.3.
Due to the restriction that we impose of using real (as opposite to imaginary and complex) quantities exclusively in this book, all arguments of real-valued functions which are
not defined for negative quantities, like square roots and logarithmic functions, are assumed to be non-negative by taking the absolute value, if necessary, without using the
absolute value symbol. This is to simplify p
the notation and avoid potential confusion with
√
the determinant notation. So, g means |g| and ln(g) means ln(|g|).
We follow the summation convention which is widely used in the literature of tensor
calculus and its applications. However, the summation symbol (i.e. Σ) is used in a
few cases where a summation operation is needed but the conditions of the summation
convention do not apply or there is an ambiguity about them, e.g. when an index is
repeated more than twice or when a summation index is not repeated visually because it is
part of a squared symbol. In a few other cases, where a twice-repeated index that complies
with the conditions of the summation convention does not imply summation and hence
the summation convention do not apply, we clarified the situation by adding comments
like “no sum on index”. We may also add a “no sum” comment in some cases where the
conditions of the summation convention do not apply technically but the expression may
be misleading since it contains a repetitive index, e.g. when both indices are of the same
variance type in a general coordinate system such as g ii or when one of the apparent
indices is in fact a label for a scalar rather than a variable index such as |Ei | or hj .
All the transformation equations in the present book are continuous and real, and all the
derivatives are continuous over their intended domain. Based on the well known continuity
condition of differential calculus, the individual differential operators in the second (and
higher) order partial derivatives with respect to different indices are commutative, that is:
∂i ∂j = ∂j ∂i
(6)
We generally assume that this continuity condition is satisfied and hence the order of
the partial differential operators in these mixed second order partial derivatives does not
matter.
We use vertical bars (i.e. |::|) to symbolize determinants and square brackets (i.e. [::]) to
symbolize matrices. This applies when these symbols contain arrays of objects; otherwise
they have their normal meaning according to the context, e.g. bars embracing a vector
1.2 General Background about Tensors
12
such as |v| mean modulus of the vector. Also, we use indexed square brackets (such as [A]i
and [∇f ]i ) to denote the ith component of vectors in their symbolic or vector notation.
For tensors of higher rank, more than one index are used to denote their components, e.g.
[A]ij represents the ij th component of the rank-2 tensor A.
We finally should remark that although we generally talk about nD spaces, our main
focus is the low dimensionality spaces (mostly 2D and 3D) especially with regard to coordinate systems and hence some of the statements may apply only to these low dimensionality
spaces although the statements are given in the context of nD spaces. In most cases, such
statements can be generalized simply by adding extra conditions or by a slight modification
to the phrasing and terminology.
1.2
General Background about Tensors
A tensor is an array of mathematical objects (usually numbers or functions) which transforms according to certain rules under coordinates change. In an nD space, a tensor of
rank-k has nk components which may be specified with reference to a given coordinate
system. Accordingly, a scalar, such as temperature, is a rank-0 tensor with (assuming a
3D space) 30 = 1 component, a vector, such as force, is a rank-1 tensor with 31 = 3 components, and stress is a rank-2 tensor with 32 = 9 components. In Figure 1 we graphically
illustrate the structure of a rank-3 tensor in a 3D space.
The nk components of a rank-k tensor in an nD space are identified by k distinct integer indices (e.g. i, j, k) which are attached, according to the commonly-employed tensor
notation, as superscripts or subscripts or a mix of these to the right side of the symbol
utilized to label the tensor, e.g. Aijk , Aijk and Ajk
i . Each tensor index takes all the values
over a predefined range of dimensions such as 1 to n in the above example of an nD space.
In general, all tensor indices have the same range, i.e. they are uniformly dimensioned.[2]
When the range of tensor indices is not stated explicitly, it is usually assumed to range
over the values 1, 2, 3. However, the range must be stated explicitly or implicitly to avoid
ambiguity.
The characteristic property of tensors is that they satisfy the principle of invariance
under certain coordinate transformations. Therefore, formulating the fundamental laws
of physics in a tensor form ensures that they are form-invariant, and hence they are
objectively representing the physical reality and do not depend on the observer and his
coordinate system. Having the same form in different coordinate systems may also be
labeled as being covariant although this term is usually used for a different meaning in
tensor calculus, as will be explained in § 3.1.1.
While tensors of rank-0 are generally represented in a common form of light face nonindexed italic symbols like f and h, tensors of rank ≥ 1 are represented in several forms
and notations, the main ones are the index-free notation, which may also be called the
direct or symbolic or Gibbs notation, and the indicial notation which is also called the
[2]
This applies to the common cases of tensor applications, but there are instances (e.g. in differential
geometry of curves and surfaces) of tensors which are not uniformly dimensioned because the tensor is
related to two spaces with different dimensions such as a 2D surface embedded in a 3D space.
1.2 General Background about Tensors
13
A332
A331
A321
A311
A232
A231
A221
A222
A211
A131
A121
A111
A132
A122
A112
A212
A322
A312
A333
A323
A313
A233
A223
A213
A133
A123
A113
Figure 1: Graphical illustration of a rank-3 tensor Aijk in a 3D space, i.e. each one of
i, j, k ranges over 1, 2, 3.
index or component or tensor notation. The first is a geometrically oriented notation with
no reference to a particular coordinate system and hence it is intrinsically invariant to the
choice of coordinate systems, whereas the second takes an algebraic form based on components identified by indices and hence the notation is suggestive of an underlying coordinate
system, although being a tensor makes it form-invariant under certain coordinate transformations and therefore it possesses certain invariant properties. The index-free notation
is usually identified by using bold face non-italic symbols, like a and B, while the indicial
notation is identified by using light face indexed italic symbols such as ai and Bij . It is
noteworthy that although rank-0 and rank-1 tensors are, respectively, scalars and vectors,
not all scalars and vectors (in their generic sense) are tensors of these ranks. Similarly,
rank-2 tensors are normally represented by square matrices but not all square matrices
represent rank-2 tensors.
1.2 General Background about Tensors
14
Tensors can be combined through common algebraic operations such as addition and
multiplication. Tensor term is a product of tensors including scalars and vectors and may
consist of a single tensor which can be regarded as a multiple of unity. Tensor expression
is an algebraic sum of tensor terms which may be a trivial sum in the case of a single
term. Tensor equality is an equality of two tensor terms and/or expressions. An index
that occurs once in a tensor term is a free index while an index that occurs twice in a
tensor term is a dummy or bound index.
The order of a tensor is identified by the number of its indices. For example, Aijk is
a tensor of order 3 while Bkm is a tensor of order 2. The order of the tensor normally
identifies its rank as well and hence Aijk is of rank-3 and Bkm is of rank-2. However, when
the operation of contraction of indices (see § 3.2.4) takes place once or more, the order of
the tensor is not affected but its rank is reduced by two for each contraction operation.
Hence, the order of a tensor is equal to the number of all of its indices including the dummy
indices, while the rank is equal to the number of its free indices only. Accordingly, Aab
abmn
a
is of order 6 and rank-2 while Bai
is of order 3 and rank-1. We note that many authors
follow different conventions such as using “order” as equivalent to what we call “rank”.
Tensors whose all indices are subscripts, like Aij , are called covariant, while tensors whose
all indices are superscripts, like Ak , are called contravariant. Tensors with both types of
indices, like Aamn
ak , are called mixed type. Subscript indices, rather than subscripted tensors, are also described as covariant and superscript indices are described as contravariant.
The zero tensor is a tensor whose all components are zero. The unit tensor or unity tensor,
which is usually defined for rank-2 tensors, is a tensor whose all elements are zero except
those with identical values of all indices (e.g. A11 or B 33 ) which are assigned the value 1.
There are general rules that govern the manipulation of tensors and hence they should be
observed in the handling of mathematical expressions and calculations of tensor calculus.
One of these rules is that no tensor index is allowed to occur more than twice in a legitimate
tensor term. However, we follow in this assertion the common literature of tensor calculus
which represents the ordinary use of repeated indices in tensor terms. In fact, there are
many instances in the literature of tensor calculus where indices are legitimately repeated
more than twice in a single term. The bottom line is that as long as the tensor expression
makes sense and the intention is clear, such repetitions should be allowed with no need to
take special precautions like using parentheses as done by some authors. In particular, the
forthcoming summation convention will not apply automatically in such cases although
summation on such indices, if needed, can be carried out explicitly, by using the summation
symbol Σ, or by a special declaration of such intention similar to the summation convention
which is usually restricted to the twice-repeated indices.
Regarding the aforementioned summation convention, according to this convention which
is widely used in the literature of tensor calculus including the present book, dummy
indices imply summation over their range. More clearly, a twice-repeated variable (i.e.
not numeric) index in a single term implies a sum of terms equal in number to the range
of the repeated index. Hence, in a 3D space we have:
1j
2j
3j
Aaj
a = A1 + A2 + A3
(7)
1.2 General Background about Tensors
15
while in a 4D space we have:
Bi + Cia Da = Bi + Ci1 D1 + Ci2 D2 + Ci3 D3 + Ci4 D4
(8)
We note that although the twice-repeated index should be in the same term for the summation convention to apply, it does not matter if the two indices occur in one tensor or in
two tensors, as seen in the last example where the summation convention applies to a.
We should also remark that there are many cases in the mathematics of tensors where
a repeated index is needed but with no intention of summation, such as using Aii or Ajj
to mean the components of the tensor A whose indices have identical value like A11 and
A22 . So to avoid confusion, when the dummy indices in a particular case do not imply
summation, the situation must be clarified by enclosing such indices in parentheses or by
underscoring or by using upper case letters with declaration of these conventions, or by
adding a clarifying comment like “no summation over repeated indices”. These precautions
are obviously needed if the summation convention is adopted in general but it does not
apply in some exceptional cases where repeated indices are needed in the notation with
no implication of summation.
Another rule of tensors is that a free index should be understood to vary over its range
(e.g. 1, . . . , n) which is determined by the space dimension and hence it should be interpreted as saying “for all components represented by the index”. Therefore, a free index
represents a number of terms or expressions or equalities equal to the number of the allowed values of its range. For example, when i and j can vary over the range 1, . . . , n then
the expression Ai + Bi represents n separate expressions while the equation Aji = Bij represents n × n separate equations which represent the combination of all possible n values
of i with all possible n values of j, that is:
A1 + B1 , A2 + B2 , · · · , An + Bn
(9)
A11 = B11 , A21 = B12 , A12 = B21 , · · · , Ann = Bnn
(10)
Also, each tensor index should conform to one of the forthcoming variance transformation rules as given by Eqs. 71 and 72, i.e. it is either covariant or contravariant. For
orthonormal Cartesian coordinate systems, the two variance types (i.e. covariant and contravariant) do not differ because the metric tensor is given by the Kronecker delta (refer
to § 4.1 and 4.5) and hence any index can be upper or lower although it is common to use
lower indices in this case. We note that orthonormal vectors mean a set of vectors which
are mutually orthogonal and each one is of unit length, while orthonormal coordinate
system means a coordinate system whose basis vector set is orthonormal at all points of
the space where the system is defined (see § 2.6). The orthonormality of vectors may be
expressed mathematically by:
Vi · Vj = δij
or
Vi · Vj = δ ij
(11)
where the indexed δ is the Kronecker delta symbol and the indexed V symbolizes a vector
in the set.
1.2 General Background about Tensors
16
In this context, we should remark that for tensor invariance, a pair of dummy indices
involved in summation should in general be complementary in their variance type, i.e.
one covariant and the other contravariant. However, for orthonormal Cartesian systems
the two variance types are the same and hence when both dummy indices are covariant
or both are contravariant it should be understood as an indication that the underlying
coordinate system is orthonormal Cartesian if the possibility of an error is excluded.
As indicated earlier, tensor order is equal to the number of its indices while tensor rank is
equal to the number of its free indices. Hence, scalars (terms, expressions and equalities)
have no free index since they are of rank-0, and vectors have a single free index while
rank-2 tensors have exactly two free indices. Similarly, rank-n tensors have exactly n
free indices. The dimension of a tensor is determined by the range taken by its indices
which represents the number of dimensions of the underlying space. For example, in a 3D
space the tensor Aji is of rank-2 because it possesses exactly two free indices but it is of
dimension three since each one of its free indices range over the values 1, 2, 3 and hence it
may be represented by a 3 × 3 matrix. However, in a 4D space this rank-2 tensor will be
represented by a 4 × 4 matrix since its two free indices range over 1, 2, 3, 4.
The rank of all terms in legitimate tensor expressions and equalities must be the same
and hence:
Aji − Bij
and
Aji = Cij
(12)
are legitimate but:
Aji − Bi
and
jk
Ajk
i − Bi
and
j
Ajim = Cim
(14)
jm
Ajk
i − Bi
and
j
Ajim = Cin
(15)
Aji = C j
(13)
are illegitimate. Moreover, each term in valid tensor expressions and equalities must have
the same set of free indices (e.g. i, j, k) and hence:
are legitimate but:
are illegitimate although they are all of the same rank.
Also, a free index should keep its variance type in every term in valid tensor expressions
and equations, i.e. it must be covariant in all terms or contravariant in all terms, and
hence:
mk
Amk
and
Aji = Cij
(16)
iq − Biq
are legitimate but:
Aji − Bji
and
Ajin = Cijn
(17)
are illegitimate although they are all of the same rank and have the same set of free
indices. We also note that the order of the tensor indices in legitimate tensor expressions
and equalities should be the same and hence:
Akm + Ckm
and
Djin = E jin
(18)
1.2 General Background about Tensors
17
are legitimate but:
i
Aijk + Ckj
and
Dinm = Eimn
(19)
are illegitimate although they are all of the same rank, have the same set of free indices
and have the same variance type.
We remark that expressions and equalities like:
Aij + Bji
and
iq
qi
Aiq
k = B k − Ck
(20)
are common in the literature of tensor calculus and its applications which may suggest that
the order of indices (or another one of the aforementioned features of the indicial structure)
in tensor expressions and equalities is not important. However, expressions and equalities
like these refer to the individual components of these tensors and hence they are of scalar,
rather than tensor, nature. Hence the expression Aij + Bji means adding the value of the
component Aij of tensor A to the value of the component Bji of tensor B and not adding
iq
qi
the tensor A to the tensor B. Similarly, the equality Aiq
k = Bk − Ck means subtracting
the value of the component Ckiq of tensor C from the value of the component Bkqi of tensor
B to obtain the value of the component Aiq
k of tensor A. We may similarly write things
i
like Ai = B or Ai = Bj to mean the equality of the values of these components and not
tensor equality.
As we will see (refer to § 3.1.1), the indicial notation of tensors is made with reference to
a set of basis vectors. For example, when we write Aijk as a tensor we mean Aijk Ei Ej Ek .
This justifies all the above rules about the indicial structure (rank, set of free indices,
variance type and order of indices) of tensor terms involved in tensor expressions and
equalities because this structure is based on a set of basis vectors in a certain order.
Therefore, an expression like Aijk + B ijk means Aijk Ei Ej Ek + B ijk Ei Ej Ek and an equation
like Aijk = B ijk means Aijk Ei Ej Ek = B ijk Ei Ej Ek . However, as seen above it is common
in the literature of tensor calculus that the tensor notation like Aijk is used to label the
components and hence the above rules are not respected (e.g. Bij +Cji or ǫij = ǫij ) because
these components are scalars in nature and hence these expressions and equalities do not
refer to the vector basis. In this context, we should remark that an additional condition
may be imposed on the indicial structure that is all the indices of a tensor should refer to
a single vector basis set and hence a tensor cannot be indexed in reference to two different
basis sets simultaneously.[3]
While free indices should be named uniformly in all terms of tensor expressions and
equalities, dummy indices can be named in each term independently and hence:
b
c
Aaak + Bbk
+ Cck
[3]
and
ja
Dij = Eia
+ Fibjb
(21)
A tensor of different dimensionality (as in differential geometry) may not be compliant with this condition and hence we may add an extra condition that all the indices should refer to a single vector basis
set for any particular space. Alternatively, a tensor like this can be seen as a tensor in reference to each
individual type of indices separately but it is not a tensor simultaneously to both types of indices where
the simultaneously condition takes care of this.
1.2 General Background about Tensors
18
are legitimate. The reason is that a dummy index represents a sum in its own term with
no reach or presence into other terms. Despite the above restriction on the free indices, a
free index in an expression or equality can be renamed uniformly and thoroughly using a
different symbol, as long as this symbol is not already in use, assuming that both symbols
vary over the same range, i.e. they have the same dimension. For example, we can change:
jk
Ajk
i + Bi
to
jk
Ajk
p + Bp
(22)
j
j
− Dim
Ajim = Cim
to
j
j
Ajpm = Cpm
− Dpm
(23)
and change:
as long as i and p have the same range and p is not already in use as an index for
another purpose in that context. As indicated, the change should be thorough and hence
all occurrences of the index i in that context, which may include other expressions and
equalities, should be subject to that change. Regarding the dummy indices, they can be
replaced by another symbol which is not present (as a free or dummy index) in their term
as long as there is no confusion with a similar symbol in that context.
Indexing is generally distributive over the terms of tensor expressions and equalities. For
example, we have:
[A + B]i = [A]i + [B]i
(24)
and
[A = B]i
⇐⇒
[A]i = [B]i
(25)
Unlike scalars and tensor components, which are essentially scalars in a generic sense,
operators cannot in general be freely reordered in tensor terms. Therefore, we have the
following legitimate equalities:
f h = hf
and
Ai B j = B j Ai
(26)
but we cannot equate ∂i Aj to Aj ∂i since in general we have:
∂i Aj 6= Aj ∂i
(27)
This should be obvious because ∂i Aj means that ∂i is operating on Aj but Aj ∂i means
that ∂i is operating on something else and Aj just multiplies the result of this operation.
As seen above, the order of the indices[4] of a given tensor is important and hence it
should be observed and clarified, because two tensors with the same set of free indices and
with the same indicial structure that satisfies the aforementioned rules but with different
indicial order are not equal in general. For example, Aijk is not equal to Ajik unless A
is symmetric with respect to the indices i and j (refer to § 3.1.5). Similarly, B mln is not
equal to B lmn unless B is symmetric in its indices l and m. The confusion about the order
of indices occurs specifically in the case of mixed type tensors such as Aijk which may not
[4]
The “order” here means “arrangement” and hence it should not be confused with the order of tensor as
defined above in the same context as tensor rank.
1.3 Exercises and Revision
19
be clear since the order can be ijk or jik or jki. Spaces are usually used in this case to
clarify the order. For example, the latter tensor is symbolized as Ai jk if the order of the
indices is ijk, and as Aj i k if the order of the indices is jik while it is symbolized as Ajk i
if the order of the indices is jki. Dots may also be used in such cases to indicate, more
explicitly, the order of the indices and remove any ambiguity. For example, if the indices
i, j, k of the tensor A, which is covariant in i and k and contravariant in j, are of that
order, then A may be symbolized as Ai j. k where the dot between i and k indicates that j
is in the middle.
We note that in many places in this book (like many other books of tensor calculus) and
mostly for the sake of convenience in typesetting, the order of the indices of mixed type
tensors is not clarified by spacing or by inserting dots. This commonly occurs when the
order of the indices is irrelevant in the given context (e.g. any order satisfies the intended
purpose) or when the order is clear. Sometimes, the order of the indices may be indicated
. jk
implicitly by the alphabetical order of the selected indices, e.g. writing Ajk
i to mean Ai .
Finally, scalars, vectors and tensors may be defined on a single point of the space or on
a set of separate points. They may also be defined over an extended continuous region
(or regions) of the space. In the latter case we have scalar fields, vector fields and tensor
fields, e.g. temperature field, velocity field and stress field respectively. Hence, a “field”
is a function of coordinates which is defined over a given region of the space. As stated
earlier, “tensor” may be used in a general sense to include scalar and vector and hence
“tensor field” may include all the three types.
1.3
Exercises and Revision
1.1 Differentiate between the symbols used to label scalars, vectors and tensors of rank
> 1.
1.2 What the comma and semicolon in Ajk,i and Ak;i mean?
1.3 State the summation convention and explain its conditions. To what type of indices
this convention applies?
1.4 What is the number of components of a rank-3 tensor in a 4D space?
1.5 A symbol like Bijk may be used to represent tensor or its components. What is the
difference between these two representations? Do the rules of indices apply to both
representations or not? Justify your answer.
1.6 What is the meaning of the following symbols: ∇, ∂j , ∂kk , ∇2 , ∂φ , h,jk , Ai;n , ∂ n , ∇;k
and Ci;km ?
1.7 What is the difference between symbolic notation and indicial notation? For what
type of tensors these notations are used? What are the other names given to these
types of notation?
1.8 “The characteristic property of tensors is that they satisfy the principle of invariance
under certain coordinate transformations”. Does this mean that the components of
tensors are constant? Why this principle is very important in physical sciences?
1.9 State and explain all the notations used to represent tensors of all ranks (rank-0,
rank-1, rank-2, etc.). What are the advantages and disadvantages of using each one
1.3 Exercises and Revision
20
of these notations?
1.10 State the continuity condition that should be met if the equality: ∂i ∂j = ∂j ∂i is to be
correct.
1.11 Explain the difference between free and bound tensor indices. Also, state the rules
that govern each one of these types of index in tensor terms, expressions and equalities.
1.12 Explain the difference between the order and the rank of tensors and link this to the
free and dummy indices.
1.13 What is the difference between covariant, contravariant and mixed type tensors? Give
an example for each.
1.14 What is the meaning of “unit” and “zero” tensors? What is the characteristic feature
of these tensors with regard to the value of their components?
1.15 What is the meaning of “orthonormal vector set” and “orthonormal coordinate system”? State any relevant mathematical condition.
1.16 What is the rule that governs the pair of dummy indices involved in summation
regarding their variance type in general coordinate systems? Which type of coordinate
system is exempt of this rule and why?
1.17 State all the rules that govern the indicial structure of tensors involved in tensor
expressions and equalities (rank, set of free indices, variance type and order of indices).
1.18 How many equalities that the following equation contains assuming a 4D space: Bik =
Cik ? Write all these equalities explicitly, i.e. B11 = C11 , B12 = C12 , etc.
1.19 Which of the following tensor expressions is legitimate and which is not, giving detailed
explanation in each case?
Aki − Bi ,
n
− Bbb ,
Caa + Dm
a + B,
abj
cdj
+ Fabk
Scdk
1.20 Which of the following tensor equalities is legitimate and which is not, giving detailed
explanation in each case?
A.i n = B.ni ,
ab
D = Scc + Nba
,
3a + 2b = Jaa ,
k
Bkm = Cm
,
Bj = 3c − Dj
1.21 Explain why the indicial structure (rank, set of free indices, variance type and order of
indices) of tensors involved in tensor expressions and equalities are important referring
in your explanation to the vector basis set to which the tensors are referred. Also
explain why these rules are not observed in the expressions and equalities of tensor
components.
1.22 Why free indices should be named uniformly in all terms of tensor expressions and
equalities while dummy indices can be named in each term independently?
1.23 What are the rules that should be observed when replacing the symbol of a free index
with another symbol? What about replacing the symbols of dummy indices?
1.24 Why in general we have: ∂i Aj 6= Aj ∂i ? What are the situations under which the
following equality is valid: ∂i Aj = Aj ∂i ?
1.25 What is the difference between the order of a tensor and the order of its indices?
1.26 In which case Aijk is equal to Aikj ? What about Aijk and Aikj ?
1.27 What are the rank, order and dimension of the tensor Aijk in a 3D space? What about
the scalar f and the tensor Aabm
abjn from the same perspectives?
1.3 Exercises and Revision
21
1.28 What is the order of indices in Aj i k ? Insert a dot in this symbol to make the order
more explicit.
1.29 Why the order of indices of mixed tensors may not be clarified by using spaces or
inserting dots?
1.30 What is the meaning of “tensor field”? Is Ai a tensor field considering the spatial
dependency of Ai and the meaning of “tensor”?
Chapter 2
Spaces, Coordinate Systems and Transformations
The focus of this chapter is coordinate systems, their types and transformations as well as
some general properties of spaces which are needed for the development of the concepts
and techniques of tensor calculus in the present and forthcoming chapters. The chapter
also includes other sections which are intimately linked to these topics.
2.1
Spaces
A Riemannian space is a manifold characterized by the existing of a symmetric rank-2
tensor called the metric tensor. The components of this tensor, which can be in covariant
form gij or contravariant form g ij , as well as mixed form gji , are continuous variable
functions of coordinates in general, that is:
gij = gij (u1 , u2 , . . . , un )
g ij = g ij (u1 , u2 , . . . , un )
gji = gji (u1 , u2 , . . . , un )
(28)
(29)
(30)
where the indexed u symbolizes general coordinates. This tensor facilitates, among other
things, the generalization of the concept of length in general coordinate systems where the
length of an infinitesimal element of arc, ds, is defined by:
(ds)2 = gij dui duj
(31)
In the special case of a Euclidean space coordinated by an orthonormal Cartesian system,
the metric becomes the identity tensor, that is:
gij = δij
g ij = δ ij
gji = δji
(32)
More details about the metric tensor and its significance and roles will be given in § 4.5.
The metric of a Riemannian space may be called the Riemannian metric. Similarly, the
geometry of the space may be described as the Riemannian geometry. All spaces dealt
with in the present book are Riemannian with well-defined metrics. As we will see, an
nD manifold is Euclidean iff the Riemann-Christoffel curvature tensor vanishes identically
(see § 7.2.1); otherwise the manifold is curved to which the general Riemannian geometry
applies. In metric spaces, the physical quantities are independent of the form of description, being covariant or contravariant, as the metric tensor facilitates the transformation
between the different forms; hence making the description objective.
22
2.1 Spaces
23
A manifold, such as a 2D surface or a 3D space, is called “flat” if it is possible to find
a coordinate system for the manifold with a diagonal metric tensor whose all diagonal
elements are ±1; the space is called “curved” otherwise. More formally, an nD space is
described as flat space iff it is possible to find a coordinate system for which the length
of an infinitesimal element of arc ds is given by:
2
1 2
2 2
n 2
(ds) = ζ1 (du ) + ζ2 (du ) + . . . + ζn (du ) =
n
X
ζi (dui )2
(33)
i=1
where the indexed ζ are ±1 while the indexed u are the coordinates of the space. For the
space to be flat (i.e. globally not just locally), the condition given by Eq. 33 should apply
all over the space and not just at certain points or regions.
An example of flat space is the 3D Euclidean space which can be coordinated by an
orthonormal Cartesian system whose metric tensor is diagonal with all the diagonal elements being +1. This also applies to plane surfaces which are 2D flat spaces that can be
coordinated by 2D orthonormal Cartesian systems. Another example is the 4D Minkowski
space-time manifold associated with the mechanics of Lorentz transformations whose metric is diagonal with elements of ±1 (see Eq. 241). When all the diagonal elements of
the metric tensor of a flat space are +1, the space and the coordinate system may be described as homogeneous (see § 2.2.3). All 1D spaces are Euclidean and hence they cannot
be curved intrinsically, so twisted curves are curved only when viewed externally from the
embedding space which they reside in, e.g. the 2D space of a surface curve or the 3D space
of a space curve. This is because any curve can be mapped isometrically to a straight line
where both are naturally parameterized by arc length. An example of curved space is the
2D surface of a sphere or an ellipsoid since there is no possibility of coordinating these
spaces with valid 2D coordinate systems that satisfy the above criterion.
A curved space may have constant curvature all over the space, or have variable curvature
and hence the curvature is position dependent. An example of a space of constant curvature
is the surface of a sphere of radius R whose curvature (i.e. Riemannian curvature) is R12
at each point of the surface. Torus and ellipsoid are simple examples of 2D spaces with
variable curvature. Schur theorem related to nD spaces (n > 2) of constant curvature
states that: if the Riemann-Christoffel curvature tensor (see § 7.2.1) at each point of a
space is a function of the coordinates only, then the curvature is constant all over the
space. Schur theorem may also be stated as: the Riemannian curvature is constant over
an isotropic region of an nD (n > 2) Riemannian space.
A necessary and sufficient condition for an nD space to be intrinsically flat is that the
Riemann-Christoffel curvature tensor of the space vanishes identically. Hence, cylinders
are intrinsically flat, since their Riemann-Christoffel curvature tensor vanishes identically,
although they are curved as seen extrinsically from the embedding 3D space. On the other
hand, planes are intrinsically and extrinsically flat. In brief, a space is intrinsically flat
iff the Riemann-Christoffel curvature tensor vanishes identically over the space, and it is
extrinsically (as well as intrinsically) flat iff the curvature tensor vanishes identically over
the whole space. This is because the Riemann-Christoffel curvature tensor characterizes
2.2 Coordinate Systems
24
the space curvature from an intrinsic perspective while the curvature tensor characterizes
the space curvature from an extrinsic perspective.
As indicated above, the geometry of curved spaces is usually described as the Riemannian
geometry. One approach for investigating the Riemannian geometry of a curved manifold
is to embed the manifold in a Euclidean space of higher dimensionality and inspect the
properties of the manifold from this perspective. This approach is largely followed, for
example, in the differential geometry of surfaces where the geometry of curved 2D spaces
(twisted surfaces) is investigated by immersing the surfaces in a 3D Euclidean space and
examining their properties as viewed from this external enveloping 3D space. Such an
external view is necessary for examining the extrinsic geometry of the space but not its
intrinsic geometry. A similar approach may also be followed in the investigation of surface
and space curves.
2.2
Coordinate Systems
In simple terms, a coordinate system is a mathematical device, essentially of geometric
nature, used by an observer to identify the location of points and objects and describe
events in generalized space which may include space-time. In tensor calculus, a coordinate
system is needed to define non-scalar tensors in a specific form and identify their components in reference to the basis set of the system. An nD space requires a coordinate
system with n mutually independent variable coordinates to be fully described so that any
point in the space can be uniquely identified by the coordinate system. We note that the
coordinates are generally real quantities although this may not apply in some cases (see §
2.2.3).
As we will see in § 2.4, coordinate systems of 3D spaces are characterized by having
coordinate curves and coordinate surfaces where the coordinate curves occur at the intersection of the coordinate surfaces.[5] The coordinate curves represent the curves along
which exactly one coordinate varies while the other coordinates are held constant. Conversely, the coordinate surfaces represent the surfaces over which exactly one coordinate is
held constant while the other coordinates vary. At any point P in a 3D space coordinated
by a 3D coordinate system, we have 3 independent coordinate curves and 3 independent
coordinate surfaces passing through P . The 3 coordinate curves uniquely identify the
set of 3 mutually independent covariant basis vectors at P . Similarly, the 3 coordinate
surfaces uniquely identify the set of 3 mutually independent contravariant basis vectors at
P . Further details about this issue will follow in § 2.4.
There are many types and categories of coordinate system; some of which will be briefly
investigated in the following subsections. The most commonly used coordinate systems are:
Cartesian, cylindrical and spherical. The most universal type of coordinate system is the
general coordinate system which can include any type (rectilinear, curvilinear, orthogonal,
etc.). A subset of the general coordinate system is the orthogonal coordinate system which
is characterized by having mutually perpendicular coordinate curves, as well as mutually
[5]
In fact, these concepts can be generalized to nD spaces by generalizing the concepts of curves and
surfaces. However, the main interest here and in the forthcoming sections is 3D spaces.
2.2.1 Rectilinear and Curvilinear Coordinate Systems
25
perpendicular coordinate surfaces, at each point in the region of space over which the
system is defined and hence its basis vectors, whether covariant or contravariant, are
mutually perpendicular.
The coordinates of a system can have the same physical dimension or different physical
dimensions. An example of the first is the Cartesian coordinate system, which is usually identified by (x, y, z), where all the coordinates have the dimension of length, while
examples of the second include the cylindrical and spherical systems, which are usually
identified by (ρ, φ, z) and (r, θ, φ) respectively, where some coordinates, like ρ and r, have
the dimension of length while other coordinates, like φ and θ, are dimensionless. We also
note that the physical dimensions of the components and basis vectors of the covariant
and contravariant forms of a tensor are generally different.
In the following subsections we outline a number of general types and categories of
coordinate systems based on different classifying criteria. These categories are generally
overlapping and may not be exhaustive in their domain.
2.2.1
Rectilinear and Curvilinear Coordinate Systems
Rectilinear coordinate systems are characterized by the property that all their coordinate
curves are straight lines and all their coordinate surfaces are planes, while curvilinear
coordinate systems are characterized by the property that at least some of their coordinate
curves are not straight lines and some of their coordinate surfaces are not planes (see §
2.4). Consequently, the basis vectors of rectilinear systems are constant while the basis
vectors of curvilinear systems are variable in general since their direction or/and magnitude
depend on the position in the space and hence they are coordinate dependent.
Rectilinear coordinate systems can be rectangular (or orthogonal) when their coordinate
curves, as well as their coordinate surfaces, are mutually orthogonal such as the well
known rectangular Cartesian system. They can also be oblique when at least some of
their coordinate curves and coordinate surfaces do not satisfy this condition. Figure 2 is
a simple graphic illustration of rectangular and oblique rectilinear coordinate systems in
a 3D space. Similarly, curvilinear coordinate systems can be orthogonal, when the vectors
in their covariant or contravariant basis set are mutually orthogonal at each point in the
space, and can be non-orthogonal when this condition is not met. Rectilinear coordinate
systems may also be labeled as affine or linear coordinate systems although the terminology
is not universal and hence these labels may be used differently.
As stated above, curvilinear coordinate systems are characterized by the property that
at least some of their coordinate curves are not straight lines and some of their coordinate
surfaces are not planes (see Figures 3 and 10). This means that some (but not all) of
the coordinate curves of curvilinear coordinate systems can be straight lines and some
(but not all) of their coordinate surfaces can be planes. This is the case in the cylindrical
and spherical coordinate systems as we will see next. Also, the coordinate curves of
curvilinear coordinate systems may be regularly shaped curves such as circles and may be
irregularly shaped and hence they are generalized twisted curves. Similarly, the coordinate
surfaces of curvilinear systems may be regularly shaped surfaces such as spheres and may
2.2.2 Orthogonal Coordinate Systems
26
z
z
O
y
y
x
(a)
O
x
(b)
Figure 2: The two main types of rectilinear coordinate systems in 3D spaces: (a) rectangular and (b) oblique.
be irregularly shaped.
Prominent examples of curvilinear coordinate systems are the cylindrical and spherical
systems of 3D spaces. All the coordinate curves and coordinate surfaces of these systems
are regularly shaped. As we will see (refer to § 2.4), in the cylindrical coordinate systems
the ρ, φ, z coordinate curves are straight lines, circles and straight lines respectively, while
the ρ, φ, z coordinate surfaces are cylinders, semi-planes and planes respectively. Similarly,
in the spherical coordinate systems the r, θ, φ coordinate curves are straight lines, semicircles and circles respectively, while the r, θ, φ coordinate surfaces are spheres, cones and
semi-planes respectively. We note that an admissible coordinate transformation from a
rectilinear system defines another rectilinear system if the transformation is linear, and
defines a curvilinear system if the transformation is nonlinear.
2.2.2
Orthogonal Coordinate Systems
The characteristic feature of orthogonal coordinate systems, whether rectilinear or curvilinear, is that their coordinate curves, as well as their coordinate surfaces, are mutually
perpendicular at each point in their space. Hence, the vectors of their covariant basis set
and the vectors of their contravariant basis set are mutually orthogonal. As a result, the
corresponding covariant and contravariant basis vectors in orthogonal coordinate systems
have the same direction and therefore if the vectors of these basis sets are normalized they
will be identical, i.e. the normalized covariant and the normalized contravariant basis
vector sets are the same.
Prominent examples of orthogonal coordinate systems are rectangular Cartesian, cylindrical and spherical systems of 3D spaces (refer to Figure 4). A necessary and sufficient
condition for a coordinate system to be orthogonal is that its metric tensor is diagonal.
2.2.3 Homogeneous Coordinate Systems
27
u3
E3
P
u2
E2
E1
u1
Figure 3: General curvilinear coordinate system in a 3D space and its covariant basis
vectors E1 , E2 and E3 (see § 2.6) as tangents to the shown coordinate curves at a particular
point of the space P , where u1 , u2 and u3 represent general coordinates.
This can be inferred from the definition of the components of the metric tensor as dot
products of the basis vectors (see Eq. 48) since the dot product involving two different
vectors (i.e. Ei · Ej or Ei · Ej with i 6= j) will vanish if the basis vectors, whether covariant
or contravariant, are mutually perpendicular.
2.2.3
Homogeneous Coordinate Systems
When all the diagonal elements of a diagonal metric tensor of a flat space are +1, the
coordinate system is described as homogeneous. In this case the length of line element ds
of Eq. 33 becomes:
(ds)2 = dui dui
(34)
An example of homogeneous coordinate systems is the orthonormal Cartesian system of a
3D Euclidean space (Figure 4 a). A homogeneous coordinate system can be transformed
to another homogeneous coordinate system only by linear transformations. Moreover,
any coordinate system obtained from a homogeneous coordinate system by an orthogonal
transformation (see § 2.3.3) is also homogeneous. As a consequence of the last statements,
infinitely many homogeneous coordinate systems can be constructed in any flat space.
A coordinate system of a flat space can always be homogenized by allowing the coordinates to be imaginary. This is done by redefining the coordinates as:
p
(no sum over i)
(35)
U i = ζ i ui
where ζi = ±1. The new coordinates U i are real when ζi = 1 and imaginary when ζi = −1.
2.2.3 Homogeneous Coordinate Systems
28
z
e3
O
e1
e2
x
y
(a) Orthonormal Cartesian system and its orthonormal basis vectors e1 , e2 , e3
z
z
ez
e
P eρ φ
O
y
φ
ρ
x
(b) Cylindrical coordinate system and its orthonormal basis vectors eρ , eφ , ez
z
θ
O
φ
er
P eφ
r eθ
y
x
(c) Spherical coordinate system and its orthonormal basis vectors er , eθ , eφ
Figure 4: The three prominent orthogonal coordinate systems in 3D spaces: (a) orthonormal Cartesian, (b) cylindrical and (c) spherical.
2.3 Transformations
29
Consequently, the length of line element ds will be given by:
(ds)2 = dU i dU i
(36)
which is of the same form as Eq. 34. An example of a homogeneous coordinate system
with some real and some imaginary coordinates is the coordinate system of a Minkowski
4D space-time of the mechanics of Lorentz transformations. We note that homogenization
in the above sense is based on an extension of the concept of homogeneity and it is mainly
based on the definition of the length of line element.
2.3
Transformations
In general terms, a transformation from an nD space to another nD space is a correlation that maps a point from the first space (original) to a point in the second space
(transformed) where each point in the original and transformed spaces is identified by
n independent coordinates. To distinguish between the two sets of coordinates in the
two spaces, the coordinates of the points in the transformed space may be notated with
barred symbols like (ū1 , ū2 , . . . , ūn ), while the coordinates of the points in the original
space are notated with unbarred similar symbols like (u1 , u2 , . . . , un ). Under certain conditions, which will be clarified later, such a transformation is unique and hence an inverse
transformation from the transformed space to the original space is also defined.
Mathematically, each one of the direct and inverse transformations can be regarded as
a mathematical correlation expressed by a set of equations in which each coordinate in
one space is considered as a function of the coordinates in the other space. Hence, the
transformations between the two sets of coordinates in the two spaces can be expressed
mathematically in a generic form by the following two sets of independent relations:
ūi = ūi (u1 , u2 , . . . , un )
ui = ui (ū1 , ū2 , . . . , ūn )
(37)
where i = 1, 2, . . . , n with n being the space dimension. The independence of the above
relations is guaranteed iff the Jacobian of the transformation does not vanish at any point
in the space (refer to the following paragraphs about the Jacobian).
An alternative to the latter view of considering the transformation as a mapping between two different spaces is to view it as a correlation relating the same point in the same
space but observed from two different coordinate systems which are subject to a similar
transformation. The following will be largely based on the latter view although we usually
adopt the one which is more convenient in the particular context. As far as the notation
is concerned, there is no fundamental difference between the barred and unbarred systems
and hence the notation can be interchanged. We also note that the transformations considered here, and in the present book in general, are between two spaces or coordinate
systems of equal dimensions and hence we do not consider transformations between spaces
or coordinate systems of different dimensions. Consequently, the Jacobian matrix of the
transformation is always square and its determinant (i.e. the Jacobian) is defined. As
indicated earlier, if the mapping from an original rectangular Cartesian system is linear,
2.3 Transformations
30
the coordinate system obtained from such a transformation is called affine or rectilinear. Coordinate systems which are not affine are described as curvilinear although the
terminology may differ between the authors.
The following n × n matrix of n2 partial derivatives of the unbarred coordinates with
respect to the barred coordinates, where n is the space dimension, is called the “Jacobian
matrix” of the transformation between the unbarred and barred systems:
∂u1 ∂u1
1
· · · ∂∂uūn
∂ ū1
∂ ū2
∂u2 ∂u2 · · · ∂u2
1 ∂ ū2
∂ ūn
(38)
J = ∂ ū.
..
.
.
.
.
.
.
.
.
.
n
∂un
∂un
· · · ∂u
∂ ū1
∂ ū2
∂ ūn
while its determinant:
J = det(J)
(39)
is called the “Jacobian” of the transformation where the indexed u and ū are the coordinates
in the unbarred and barred coordinate systems of the nD space. The Jacobian array
(whether matrix or determinant) contains all the possible n2 partial derivatives made of
the different combinations of the u and ū indices where the pattern of the indices in this
array is simple, that is the indices of u in the numerator provide the indices for the rows
while the indices of ū in the denominator provide the indices for the columns. This labeling
scheme may be interchanged which is equivalent to taking the transpose of the array. As
it is well known, the Jacobian will not change by this transposition since the determinant
of a matrix is the same as the determinant of its transpose, i.e. det(A) = det(AT ).
We note that all coordinate transformations in the present book are continuous, single
valued and invertible. We also note that “barred” and “unbarred” in the definition of
Jacobian should be understood in a general sense not just as two labels since the Jacobian
is not restricted to transformations between two systems of the same type but labeled as
barred and unbarred. In fact the two coordinate systems can be fundamentally different in
nature such as orthonormal Cartesian and general curvilinear. The Jacobian matrix and
determinant represent any transformation by the above partial derivative array system
between two coordinate systems defined by two different sets of coordinate variables not
necessarily as barred and unbarred. The objective of defining the Jacobian as between
unbarred and barred systems is simplicity and generality.
The transformation from the unbarred coordinate system to the barred coordinate system
is bijective[6] iff J 6= 0 at any point in the transformed region of the space. In this case,
the inverse transformation from the barred to the unbarred system is also defined and
bijective and is represented by the inverse of the Jacobian matrix, that is:
J̄ = J−1
(40)
Consequently, the Jacobian of the inverse transformation, being the determinant of the
inverse Jacobian matrix, is the reciprocal of the Jacobian of the original transformation,
[6]
Bijective means the mapping is injective (one-to-one) and surjective (onto).
2.3 Transformations
31
that is:
1
(41)
J¯ =
J
As we remarked, there is no fundamental notational difference between the barred and
unbarred systems and hence the labeling is rather arbitrary and can be interchanged.
Therefore, the Jacobian may be notated as unbarred over barred or the other way around.
The essence is that the Jacobian usually represents the transformation from an original
system to another system while its inverse represents the opposite transformation although
even this is not generally respected in the literature of mathematics and hence “Jacobian”
my be used to label the opposite transformation. Yes, in a specific context when one of
these is labeled as the Jacobian, the other one should be labeled as the inverse Jacobian to
distinguish between the two opposite Jacobians and their corresponding transformations.
However, there may also be practical aspects for choosing which is the Jacobian and which
is the inverse since it is easier sometimes to compute one of these than the other and hence
we start by computing the easier as the Jacobian (whether from original to transformed
or the other way) followed by obtaining the reciprocal as the inverse Jacobian. Anyway,
in this book we generally use “Jacobian” flexibly where the context determines the nature
of the transformation.[7]
An admissible (or permissible or allowed) coordinate transformation may be defined
generically as a mapping represented by a sufficiently differentiable set of equations plus
being invertible by having a non-vanishing Jacobian (J 6= 0). More technically, a coordinate transformation is commonly described as admissible iff the transformation is
bijective with non-vanishing Jacobian and the transformation function is of class C 2 .[8]
We note that the C n continuity condition means that the function and all its first n partial
derivatives do exist and are continuous in their domain. Also, some authors may impose
a weaker continuity condition of being of class C 1 .
An object that does not change by admissible coordinate transformations is described
as “invariant” such as a true scalar (see § 3.1.2) which is characterized by its sign and
magnitude and a true vector which is characterized by its magnitude and direction in space.
Similarly, an invariant property of an object or a manifold is a property that is independent
of admissible coordinate transformations such as being form invariant which characterizes
tensors or being flat which characterize spaces under certain types of transformation. It
should be noted that an invariant object or property may be invariant with respect to
certain types of transformation but not with respect to other types of transformation and
hence the term may be used generically where the context is taken into consideration for
[7]
We may get rid of all these complications by using the term “Jacobian” to represent a system of
partial derivatives whose objective is to transform from one coordinate system to another (regardless of
anything else like system notation or being original and secondary) and hence both the “Jacobian” and
its inverse are Jacobians in this general sense; the two will be distinguished from each other by stating
from which system to which system the Jacobian transforms. Hence, we can say legitimately and with
no ambiguity: the Jacobian from A to B and the Jacobian from B to A. These may be labeled as
J (A → B) and J (B → A) with similar notation for the Jacobian matrix. This, we believe, can resolve
all these issues and avoid confusion.
[8]
The meaning of “admissible coordinate transformation” may vary depending on the context.
2.3.1 Proper and Improper Transformations
32
sensible interpretation.
A product or composition of space or coordinate transformations is a succession of transformations where the output of one transformation is taken as the input to the next transformation. For example, a series of m transformations labeled as Ti where i = 1, 2, · · · , m
may be applied sequentially onto a mathematical object O. This operation can be expressed mathematically by the following composite transformation Tc :
Tc (O) = Tm Tm−1 · · · T2 T1 (O)
(42)
where the output of T1 in this notation is taken as the input to T2 and so forth until the
output of Tm−1 is fed as an input to Tm in the end to produce the final output of Tc . In
such cases, the Jacobian of the product is the product of the Jacobians of the individual
transformations of which the product is made, that is:
Jc = Jm Jm−1 · · · J2 J1
(43)
where Ji in this notation is the Jacobian of the Ti transformation and Jc is the Jacobian
of Tc .
The collection of all admissible coordinate transformations with non-vanishing Jacobian
form a group. This means that they satisfy the properties of closure, associativity, identity and inverse. Hence, any convenient coordinate system can be chosen as the point of
entry since other systems can be reached, if needed, through the set of admissible transformations. This is one of the cornerstones of building invariant physical theories which
are independent of the subjective choice of coordinate systems and reference frames. We
remark that transformation of coordinates is not a commutative operation and hence the
result of two successive transformations may depend on the order of these transformations. This is demonstrated in Figure 5 where the composition of two rotations results in
different outcomes depending on the order of the rotations.
As there are essentially two different types of basis vectors, namely tangent vectors of
covariant nature and gradient vectors of contravariant nature (see § 2.6 and 3.1.1), there
are two main types of non-scalar tensors: contravariant tensors and covariant tensors
which are based on the type of the employed basis vectors of the given coordinate system.
Tensors of mixed type employ in their definition mixed basis vectors of the opposite type
to the corresponding indices of their components. As we will see, the transformation
between these different types is facilitated by the metric tensor of the given coordinate
system (refer to § 4.5).
In the following subsections, we briefly describe a number of types and categories of
coordinate transformations.
2.3.1
Proper and Improper Transformations
Coordinate transformations are described as “proper” when they preserve the handedness
(right- or left-handed) of the coordinate system and “improper” when they reverse the
handedness. Improper transformations involve an odd number of coordinate axes inversions in the origin of coordinates. Inversion of axes may be called improper rotation while
2.3.2 Active and Passive Transformations
33
y
z
y
z
y
x
x
z
z
y
z
x
x
z
y
x
y
x
Figure 5: The composite transformations Ry Rx (top) and Rx Ry (bottom) where Rx and
Ry represent clockwise rotation of π2 around the positive x and y axes respectively.
ordinary rotation is described as proper rotation. Figure 6 illustrates proper and improper
coordinate transformations of a rectangular Cartesian coordinate system in a 3D space.
2.3.2
Active and Passive Transformations
Transformations can be active, when they change the state of the observed object such as
rotating the object in the space, or passive when they are based on keeping the state of
the object and changing the state of the coordinate system which the object is observed
from. In brief, the subject of an active transformation is the object while the subject of a
passive transformation is the coordinate system.
2.3.3
Orthogonal Transformations
An orthogonal coordinate transformation consists of a combination of translation, rotation and reflection of axes. The Jacobian of orthogonal transformations is unity, that is
J = ±1.[9] The orthogonal transformation is described as positive iff J = +1 and neg[9]
This condition should apply even when the transformation includes a translation since the added constants that represent the translation in the transformation equations will vanish in the Jacobian matrix.
However, there seems to be a different convention that excludes translation to be part of orthogonal
transformations. There also seems to be another convention which restricts orthogonal transformations
to translation and rotation.
2.3.4 Linear and Nonlinear Transformations
34
x3
X3
x3
X3
O
X1
O
X2
x1
X1
Proper
x2
X2
x1
Improper
x2
Figure 6: Proper and improper transformations of a rectangular Cartesian coordinate
system in a 3D space where the former is achieved by a rotation of the coordinate system
while the latter is achieved by a rotation followed by a reflection of the first axis in the
origin of coordinates. The transformed systems are shown as dashed and labeled with
upper case letters while the original system is shown as solid and labeled with lower case
letters.
ative iff J = −1. Positive orthogonal transformations consist solely of translation and
rotation (possibly trivial ones as in the case of the identity transformation) while negative orthogonal transformations include reflection, by applying an odd number of axes
reversal, as well. Positive transformations can be decomposed into an infinite number of
continuously varying infinitesimal positive transformations each one of which imitates an
identity transformation. Such a decomposition is not possible in the case of negative orthogonal transformations because the shift from the identity transformation to reflection
is impossible by a continuous process.
2.3.4
Linear and Nonlinear Transformations
The characteristic property of linear transformations is that they maintain scalar multiplication and algebraic addition, while nonlinear transformations do not. Hence, if T is a
linear transformation then we have:
T (aA ± bB) = a T (A) ± b T (B)
(44)
where A and B are mathematical objects to be transformed by T and a and b are scalars.
As indicated earlier, an admissible coordinate transformation from a rectilinear system
defines another rectilinear system if the transformation is linear, and defines a curvilinear
system if the transformation is nonlinear.
2.4 Coordinate Curves and Coordinate Surfaces
2.4
35
Coordinate Curves and Coordinate Surfaces
As seen earlier, coordinate systems of 3D spaces are characterized by having coordinate
curves and coordinate surfaces where the coordinate curves represent the curves of mutual
intersection of the coordinate surfaces.[10] These coordinate curves and coordinate surfaces
play a crucial role in the formulation and development of the mathematical structures of
the coordinated space. The coordinate curves represent the curves along which exactly one
coordinate varies while the other coordinates are held constant. Conversely, the coordinate
surfaces represent the surfaces over which all coordinates vary except one which is held
constant. In brief, the ith coordinate curve is the curve along which only the ith coordinate
varies while the ith coordinate surface is the surface over which only the ith coordinate is
constant.
For example, in a 3D Cartesian system identified by the coordinates (x, y, z) the curve
r(x) = (x, c2 , c3 ), where c2 and c3 are real constants, is an x coordinate curve since x
varies while y and z are held constant, and the surface r(x, z) = (x, c2 , z) is a y coordinate
surface since x and z vary while y is held constant. Similarly, in a cylindrical coordinate
system identified by the coordinates (ρ, φ, z) the curve r(φ) = (c1 , φ, c3 ), where c1 is
a real constant, is a φ coordinate curve since ρ and z are held constant while φ varies.
Likewise, in a spherical coordinate system identified by the coordinates (r, θ, φ) the surface
r(θ, φ) = (c1 , θ, φ) is an r coordinate surface since r is held constant while θ and φ vary.
As stated before, coordinate curves represent the curves of mutual intersection of coordinate surfaces. This should be obvious since along the intersection curve of two coordinate
surfaces, where on each one of these surfaces one coordinate is held constant, two coordinates will be constant and hence only the third coordinate can vary. Hence, in a 3D
Cartesian coordinate system the x coordinate curves occur at the intersection of the y
and z coordinate surfaces, the y coordinate curves occur at the intersection of the x and
z coordinate surfaces, and the z coordinate curves occur at the intersection of the x and
y coordinate surfaces (refer to Figure 7). Similarly, in a cylindrical coordinate system the
ρ, φ and z coordinate curves occur at the intersection of the φ, z, the ρ, z and the ρ, φ
coordinate surfaces respectively (refer to Figure 8). Likewise, in a spherical coordinate
system the r, θ and φ coordinate curves occur at the intersection of the θ, φ, the r, φ and
the r, θ coordinate surfaces respectively (refer to Figure 9).
In this context, we note that the three types of coordinate surface of a Cartesian system, or in fact any rectilinear coordinate system (see § 2.2.1), are planes and hence the
three types of coordinate curves are straight lines (refer to Figure 7). As for cylindrical
coordinate systems, the ρ coordinate surfaces are cylinders (and this maybe the reason
behind labeling them as “cylindrical”), the φ coordinate surfaces are semi-planes and the z
coordinate surfaces are planes. Hence, the ρ, φ and z coordinate curves are straight lines,
circles, and straight lines respectively (refer to Figure 8). Regarding spherical coordinate
systems, the r coordinate surfaces are spheres, the θ coordinate surfaces are cones and
the φ coordinate surfaces are semi-planes. Hence, the r, θ and φ coordinate curves are
[10]
Our focus here is 3D spaces although these concepts can be generalized to nD spaces as indicated
earlier.
36
yC
S
2.5 Scale Factors
z
z CC
y CC
z CS
xC
C
x CS
y
x
Figure 7: Coordinate curves (CC) and coordinate surfaces (CS) in a 3D Cartesian coordinate system.
straight lines, semi-circles, and circles respectively (refer to Figure 9).
The coordinate surfaces of general curvilinear coordinate systems can vary in general and
hence they are not represented by particular geometric shapes such as planes and spheres.
As a result, the coordinate curves of these systems are general space curves which may not
have regular geometric shapes such as straight line or circle. As stated earlier, orthogonal
coordinate systems are characterized by having coordinate surfaces which are mutually
orthogonal at each point of the space, and consequently their coordinate curves are also
mutually orthogonal at each point of the space. Finally, we remark that the transformation
relations of Eq. 37 are used in defining the set of coordinate surfaces and coordinate curves.
2.5
Scale Factors
Scale factors of a coordinate system are those factors which are required to multiply the
coordinate differentials to obtain the distances traversed during a change in the coordinate
of that magnitude. The scale factors are symbolized with h1 , h2 , . . . , hn where this notation is usually used in orthogonal coordinate systems. For example, in the plane polar
coordinate system represented by the coordinates (ρ, φ), the scale factor of the second
coordinate φ is ρ because ρ is the factor used to multiply the differential of the polar angle
dφ to obtain the distance L traversed by a change of magnitude dφ in the polar angle,
2.5 Scale Factors
37
z
ρ CC
z CS
z
O
φ CS
φ
y
x
z
(a) ρ coordinate curvex with φ and z coordinate surfaces
z
z
z CS
z
ρ
ρ
y
φ CC
O
ρ CS
y
x
z
x
(b) φ coordinate curve with ρ zand z coordinate surfaces
z
φ CS
ρ CS
y
O
ρ
x
φ
ρ
ρ
y
y
x
z CC
x
(c) z coordinate curve with ρ and φ coordinate surfaces
Figure 8: Coordinate curves (CC) and coordinate surfaces(CS) in cylindrical coordinate
systems.
2.5 Scale Factors
38
z
θ CS
r CC
θ
O
φ
φ CS
y
x
(a) r coordinate curve with zθ and φ coordinate surfaces
z
r CS
θ CC
O
r
r
φ
x
x
y
y
φ CS
z
(b) θ coordinate curve withzr and φ coordinate surfaces
z
CS
r
θ CS
r
r
θ
y
y
y
φ CC
O
x
x
x
r CS
(c) φ coordinate curve with r and θ coordinate surfaces
Figure 9: Coordinate curves (CC) and coordinate surfaces (CS) in spherical coordinate
systems.
2.6 Basis Vectors and Their Relation to the Metric and Jacobian
39
Table 1: The scale factors (h1 , h2 , h3 ) for the three most commonly used orthogonal coordinate systems in 3D spaces: orthonormal Cartesian, cylindrical and spherical. The
squares of these entries and the reciprocals of these squares give the diagonal elements of
the covariant and contravariant metric tensors, gij and g ij , respectively of these systems
(see Eqs. 238-240).
h1
h2
h3
Cartesian (x, y, z)
1
1
1
Cylindrical (ρ, φ, z)
1
ρ
1
Spherical (r, θ, φ)
1
r
r sin θ
that is: L = ρ dφ.
The scale factors are also used for other purposes such as normalizing the basis vectors
and defining the components in the physical representation of vectors and tensors (refer
to § 2.6 and 3.3). They are also used in the analytical expressions for length, area and
volume in orthogonal coordinate systems, as described in § 4.6.8, 4.6.9 and 4.6.10. The
scale factors for the Cartesian, cylindrical and spherical coordinate systems of 3D spaces
are given in Table 1.
2.6
Basis Vectors and Their Relation to the Metric and Jacobian
The set of basis vectors in a given manifold plays a pivotal role in the theoretical construction of the geometry of the manifold, where these vectors are used in the definition
and construction of essential concepts and objects such as the metric tensor of the space.
The basis sets are defined at each regular point in the space and hence the vectors in the
basis sets vary in general from one point to another, i.e. they are coordinate dependent.
The vectors providing the basis set for a coordinate system, which are not necessarily of
unit length or mutually orthogonal, can be of covariant type or contravariant type. The
covariant basis vectors are defined as the tangent vectors to the coordinate curves, while
the contravariant basis vectors are defined as the gradient of the space coordinates and
hence they are perpendicular to the coordinate surfaces (refer to Figure 10). Formally, the
covariant and contravariant basis vectors are defined respectively by:
Ei =
∂r
∂ui
Ei = ∇ui
(45)
where r is the position vector in Cartesian coordinates (x1 , . . . , xn ), ui represents general
coordinates, n is the space dimension and i = 1, · · · , n. As indicated above, the covariant
and contravariant basis sets, Ei and Ei , in general coordinate systems are functions of
coordinates, that is:
(46)
E i = Ei u1 , . . . , u n
E i = Ei u1 , . . . , u n
Hence, the definitions of Eq. 45 apply to each individual point in the space where the
coordinate curves and coordinate surfaces belong to that particular point. For example,
2.6 Basis Vectors and Their Relation to the Metric and Jacobian
40
at any particular point P in a 3D space with a valid coordinate system we have three mutually independent coordinate curves and three mutually independent coordinate surfaces
and hence we should have three mutually independent covariant basis vectors and three
mutually independent contravariant basis vectors at P .
u3 CC
u1 CS
E3
E3
2
u CS
P
E2
E1
u2 CC
E2
E1
u1 CC
u3 CS
Figure 10: The covariant and contravariant basis vectors of a general curvilinear coordinate
system and the associated coordinate curves (CC) and coordinate surfaces (CS) at a given
point P in a 3D space.
Like other vectors, the covariant and contravariant basis vectors of a given coordinate
system are related to each other, through the metric tensor of the system, by the following
relations:
Ei = gij Ej
Ei = g ij Ej
(47)
where the metric tensor in its covariant form gij and contravariant form g ij works as an
index shifting operator to lower and raise the indices and hence change the variance type
of the basis vectors. Thus, the transformation between the covariant basis set and the
contravariant basis set of a particular coordinate system is facilitated by the metric tensor
of that system.
The basis vectors in their covariant and contravariant forms are related to the components
of the metric tensor in its covariant and contravariant forms by the following relations:
Ei · Ej = gij
Ei · Ej = g ij
(48)
2.6 Basis Vectors and Their Relation to the Metric and Jacobian
41
The covariant and contravariant basis vectors are reciprocal basis systems, and hence we
have:
Ei · Ej = δij
Ei · Ej = δ ij
(49)
where the indexed δ is the Kronecker delta tensor (refer to § 4.1). As we will see in § 4.5,
the Kronecker delta tensor represents the metric tensor in its mixed form.
In a 3D space with a right handed coordinate system the two sets of basis vectors are
linked by the following relations:
E1 =
E2 × E3
,
E1 · (E2 × E3 )
E2 =
E3 × E1
,
E1 · (E2 × E3 )
E3 =
E1 × E2
E1 · (E2 × E3 )
(50)
E3 × E1
E1 × E2
E2 × E3
,
E2 =
,
E3 =
(51)
E1 · (E2 × E3 )
E1 · (E2 × E3 )
E1 · (E2 × E3 )
The pattern in these relations is very simple, that is the covariant vectors are represented
by contravariant vectors and vice versa; moreover, the indices in the numerators on both
sides represent even permutations of 1, 2, 3.
The relations in Eqs. 50 and 51 may be expressed in a more compact form as follows:
E1 =
Ei =
Ej × Ek
Ei · (Ej × Ek )
Ei =
Ej × Ek
Ei · (Ej × Ek )
(52)
where i, j, k take respectively the values 1, 2, 3 and the other two cyclic permutations
(i.e. 2, 3, 1 and 3, 1, 2). It is worth noting that the magnitude of the scalar triple product
Ei ·(Ej × Ek ) represents the volume of the parallelepiped formedby the vectors Ei , Ej and
Ek while the magnitude of the scalar triple product Ei · Ej × Ek represents its reciprocal.
We also note that the scalar triple product is invariant to a cyclic permutation of the
symbols of the three vectors involved, i.e. Ei · (Ej × Ek ) = Ek · (Ei × Ej ) = Ej · (Ek × Ei );
in fact its magnitude is invariant even to non-cyclic permutations. Another important note
is that the magnitudes of the basis vectors in orthogonal coordinate systems are related
to the scale factors of the coordinate system by:
|Ei | = hi
Ei =
1
hi
(53)
where hi is the scale factor for the ith coordinate (see § 2.5).
Following an admissible coordinate transformation between unbarred and barred general coordinate systems, the basis vectors in these systems are related by the following
transformation rules:
∂ ūj
Ēj
∂ui
∂ui j
Ei =
Ē
∂ ūj
Ei =
∂uj
Ej
∂ ūi
∂ ūi j
E
Ēi =
∂uj
Ēi =
(54)
(55)
where the indexed u and ū represent the coordinates in the unbarred and barred systems
respectively. The transformation rules for the components can be easily deduced from
2.6 Basis Vectors and Their Relation to the Metric and Jacobian
42
the above rules. For example, for a vector A which can be represented covariantly and
contravariantly in the unbarred and barred systems as:
A = Ei Ai = Ēi Āi
(56)
A = Ei Ai = Ēi Āi
(57)
the transformation equations of its components between the two systems are given respectively by:
∂ ūj
Āj
∂ui
∂ui j
Ā
Ai =
∂ ūj
∂uj
Aj
∂ ūi
∂ ūi j
Āi =
A
∂uj
Ai =
(58)
Āi =
(59)
These transformation rules can be easily extended to higher rank tensors of different
variance types (see § 3.1.1).
For a 3D manifold with a right handed coordinate system, we have:
E1 · (E2 × E3 ) =
√
1
E1 · E2 × E3 = √
g
g
(60)
where g is the determinant of the covariant metric tensor, i.e.
(61)
g = det (gij ) = |gij |
Because Ei · Ej = gij (Eq. 48) we have:[11]
JT J = [gij ]
(62)
where J is the Jacobian matrix that transforms between Cartesian and general coordinates,
the superscript T represents matrix transposition, [gij ] is the matrix representing the
covariant metric tensor and the product on the left is a matrix product as defined in linear
algebra which is equivalent to a dot product in tensor algebra. On taking the determinant
of both sides of Eq. 62, the relation between the Jacobian and the determinant of the
metric tensor is obtained, that is:
g = J2
(63)
where J is the Jacobian of the transformation. As it should be known, the determinant
of a product is equal to the product of the determinants, that is:
det(JT J) = det(JT )det(J)
[11]
(64)
For example, for the transformation between Cartesian and general coordinate systems in 3D spaces
we have:
1
2
3
1
1
1
T
J =
∂x
∂u11
∂x
∂u12
∂x
∂u3
∂x
∂u21
∂x
∂u22
∂x
∂u3
∂x
∂u31
∂x
∂u32
∂x
∂u3
J=
∂x
∂u21
∂x
∂u31
∂x
∂u1
∂x
∂u22
∂x
∂u32
∂x
∂u2
∂x
∂u23
∂x
∂u33
∂x
∂u3
The entries of the product of these matrices then correspond to the entries of the metric tensor as
given by Eq. 216.
2.6 Basis Vectors and Their Relation to the Metric and Jacobian
43
Moreover, the determinant of a matrix is equal to the determinant of its transpose, that
is:
det(JT ) = det(J) = J
(65)
As explained earlier, in orthogonal coordinate systems the covariant and contravariant
basis vectors, Ei and Ei , at any point of the space are in the same direction, and hence the
normalization of each one of these basis sets, by dividing each basis vector by its magnitude,
produces identical orthonormal basis sets. Consequently, there is no difference between
the covariant and contravariant components of tensors with respect to such contravariant
and covariant orthonormal basis sets. This, however, is not true in general coordinate
systems where each normalized basis set is different in general from the other.
When the covariant basis vectors Ei are mutually orthogonal at each point of the space,
the following consequences will follow:
1. The contravariant basis vectors Ei are also mutually orthogonal because the corresponding vectors of each basis set are in the same direction due to the fact that the
tangent vector to the ith coordinate curve and the gradient vector of the ith coordinate surface at a given point in the space have the same direction. This may also be
established more formally by Eq. 49 since the reciprocal of an orthogonal system of
vectors should also be orthogonal. Eqs. 50 and 51 may also be used in this argument
since the cross products in the numerators are orthogonal to the multiplicands.
2. The covariant and contravariant metric tensors, gij and g ij , are diagonal with nonvanishing diagonal elements, that is:[12]
gij = 0
g ij = 0
gii 6= 0
g ii 6= 0
(i 6= j)
(no sum on i)
(66)
(67)
This can be concluded from Eq. 48 since the dot product is zero when the indices are
different due to orthogonality. Also, the dot product should be non-zero when the
indices are identical because the basis vectors cannot vanish at the regular points of
the space since the tangent to the coordinate curve and the gradient to the coordinate
surface do exist and they cannot be zero. Moreover, the metric tensor (as we will see
in § 4.5) is invertible and hence it does not vanish at any point in the space. This
guarantees that none of the diagonal elements can be zero when the metric tensor
is diagonal since the determinant, which is formed by the product of its diagonal
elements, should not vanish.
3. The diagonal elements of the covariant and contravariant metric tensors are reciprocals, that is:
1
(no summation)
(68)
g ii =
gii
This is a result from the previous point (i.e. the metric tensor is diagonal with nonvanishing diagonal elements) since the covariant and contravariant metric tensors
[12]
This also applies to the mixed type metric tensor since it is the identity tensor according to Eq. 49
although this is not restricted to orthogonal coordinate systems.
2.7 Relationship between Space, Coordinates and Metric
44
are inverses of each other. As it is well known from linear algebra, the inverse of a
diagonal matrix with non-vanishing diagonal entries is a diagonal matrix obtained
by taking the reciprocals of its diagonal elements.
4. The magnitudes of the contravariant and covariant basis vectors are reciprocals, that
is:
1
1
Ei =
=
(69)
|Ei |
hi
This result is based on the definition of vector modulus plus Eqs. 48, 68 and 53,
that is:
√
Ei = Ei · Ei
(Eq. 263)
(70)
p
(Eq. 48)
= g ii
1
=√
(Eq. 68)
gii
1
(Eq. 48)
=√
Ei · Ei
1
=
(Eq. 264)
|Ei |
1
=
(Eq. 53)
hi
with no sum over i.
The consequences that we stated in the above bullet points also follow when the contravariant basis vectors Ei are mutually orthogonal with minor amendments required in
the phrasing of the first point.
2.7
Relationship between Space, Coordinates and Metric
In this section, we try to shed some light on the relationship between space, coordinate
system and metric tensor where in this discussion we need the concepts of basis vectors
and coordinate curves and coordinate surfaces (see § 2.4 and 2.6).[13] We start from the
very abstract concept of space (say nD space) which may be visualized as a sort of vacuum
in which mathematical objects can exist. The number n represents the dimensionality of
the space which is a quantitative measure of the complexity of the space structure where
an nD space requires exactly n mutually independent variables to be fully and uniquely
described. So, any point in this nD space can be uniquely located by assigning an ordered
set of n mutually independent variables which we call “coordinates” such as u1 , · · · , un .
These coordinates vary continuously throughout the space where each coordinate ranges
over a given real interval which may be finite or infinite. All the points of the space are
then uniquely identified by varying these variables over their predefined ranges. The way
[13]
The revelation in this section is highly informal and non-rigorous with a pedagogical objective of
providing a qualitative appreciation of the relation between space, coordinate system and metric tensor
which may be useful for some novice readers.
2.7 Relationship between Space, Coordinates and Metric
45
by which the points of space are identified by the ordered sets of n coordinates may be
labeled generically as a “coordinate system” for the space.
The condition of mutual independence of the coordinate variables is important to make
the distinction between the points unique and universal and to avoid any ambiguity. Mutual independence requires that an nD space is identified by a coordinate system with
exactly n variable coordinates. As well as the condition of mutual independence, there is
another important condition for the coordinate system to be acceptable that is it should
be universal in its ability to identify all the points of the space unambiguously, and hence
the coordinate system should be thorough in its functionality with possible exclusion of a
finite number of isolated points in the space. The “coordinate system” may then be defined
more rigorously as a bijective correspondence between the set of n-tuple numbers and the
points of the space.
For the coordinate system to be of any use, the way by which the coordinates are assigned
to the individual points should be clearly defined. So, each particular way of defining the
aforementioned bijective correspondence represents a specific coordinate system associated
with the space. In fact, there are infinitely many ways of choosing the coordinate variables
and how they vary in the space, and hence there are infinitely many coordinate systems
that can be chosen to fulfill the above objectives. However, the conditions of mutual
independence of coordinates and bijectivity of mapping impose certain restrictions on the
selection of coordinates and how they can vary to achieve the intended functionalities.
As a result of the above conditions and restrictions on the coordinate variables and
how they vary to ensure their mutual independence and universal validity at each point
in the space, there are only certain allowed ways for the choice of the coordinate curves
and coordinate surfaces. So, on excluding certain types of coordinate systems which do
not meet these requirements, and hence excluding certain types of how the coordinate
curves and surfaces are defined at the individual points of the space, we are left with only
certain acceptable possibilities for identifying how the coordinates vary and hence how the
coordinate curves and surfaces are identified.
Now, since the covariant basis vectors are defined as the tangents to the coordinate
curves and the contravariant basis vectors are defined as the gradients to the coordinate
surfaces, then to each one of those acceptable possibilities there is only one possibility for
the identification of the covariant and contravariant sets of basis vectors. As soon as the
two sets of basis vectors of the selected coordinate system are identified unambiguously,
the metric tensor in its covariant and contravariant forms is uniquely identified throughout
the space through its relation to the basis vectors, i.e. Eq. 48, and hence the vagueness
about the nature of the space, its geometry and how it should be described is removed
thanks to the existence of a well defined metric tensor.
In brief, we start by defining a coordinate system which leads to the definition of coordinate curves and coordinate surfaces (and hence covariant and contravariant basis vector
sets) and this in its turn will lead to the definition of the metric tensor. In fact, this
approach may be described as procedural since it is based on the practical procedure of
defining these concepts. A more fundamental approach may start from the metric of the
space as an essential property of the space that characterizes its nature from which the
2.8 Exercises and Revision
46
basis vectors and the coordinate system are to be derived. It should be remarked that
the coordinate system and the space metric are independent entities despite the close link
between the two. The metric can be regarded as a built-in attribute that characterizes the
space but it can be demonstrated in different forms depending on the employed coordinate
system.
In this context, it is important to distinguish between the intrinsic and extrinsic properties of an nD space where the intrinsic properties refer to those properties of the space
which are seen from inside the space by an inhabitant of the space who can perceive only
the n dimensions of the space, while the extrinsic properties refer to those properties of
the space which are seen from outside the space by an inhabitant of an mD space (m > n)
that encloses the nD space where this mD inhabitant can perceive the n dimensions of the
space plus the extra (m − n) dimensions of the mD space. While the intrinsic properties
of the nD space are described and quantified by the metric of the nD space, the extrinsic
properties of the space are described and quantified by the metric of the embedding mD
space since the nD space is a subspace of the mD space and hence it is subject to the mD
metric. An example of nD and mD spaces is a 2D surface, such as a plane or a sphere,
which is embedded in a 3D Euclidean space.
2.8
Exercises and Revision
2.1 Give brief definitions to the following terms: Riemannian space, coordinate system
and metric tensor.
2.2 Discuss the main functions of the metric tensor in a given space. How many types
the metric tensor can have?
2.3 What is the meaning of “flat” and “curved” space? Give mathematical conditions for
the space to be flat in terms of the length of an infinitesimal element of arc and in
terms of the metric tensor. Why these conditions should be global for the space to be
flat?
2.4 Give common examples of flat and curved spaces of different dimensions justifying in
each case why the space is flat or curved.
2.5 Explain why all 1D spaces are Euclidean.
2.6 Give examples of spaces with constant curvature and spaces with variable curvature.
2.7 State Schur theorem outlining its significance.
2.8 What is the condition for a space to be intrinsically flat and extrinsically flat?
2.9 What is the common method of investigating the Riemannian geometry of a curved
manifold?
2.10 Give brief definitions to coordinate curves and coordinate surfaces outlining their
relations to the basis vector sets. How many independent coordinate curves and
coordinate surfaces we have at each point of a 3D space with a valid coordinate
system?
2.11 Why a coordinate system is needed in tensor formulations?
2.12 List the main types of coordinate system outlining their relations to each other.
2.13 “The coordinates of a system can have the same physical dimension or different phys-
2.8 Exercises and Revision
2.14
2.15
2.16
2.17
2.18
2.19
2.20
2.21
2.22
2.23
2.24
2.25
2.26
2.27
2.28
2.29
2.30
2.31
2.32
2.33
2.34
2.35
2.36
2.37
2.38
47
ical dimensions”. Give an example for each.
Prove that spherical coordinate systems are orthogonal.
What is the difference between rectilinear and curvilinear coordinate systems?
Give examples of common curvilinear coordinate systems explaining why they are
curvilinear.
Give an example of a commonly used curvilinear coordinate system with some of its
coordinate curves being straight lines.
Define briefly the terms “orthogonal” and “homogeneous” coordinate system.
Give examples of rectilinear and curvilinear orthogonal coordinate systems.
What is the condition of a coordinate system to be orthogonal in terms of the form
of its metric tensor? Explain why this is so.
What is the mathematical condition for a coordinate system to be homogeneous?
How can we homogenize a non-homogeneous coordinate system of a flat space?
Give examples of homogeneous and non-homogeneous coordinate systems.
Give an example of a non-homogeneous coordinate system that can be homogenized.
Describe briefly the transformation of spaces and coordinate systems stating relevant
mathematical relations.
What “injective transformation” means? Is it necessary that such a transformation
has an inverse?
Write the Jacobian matrix J of a transformation between two nD spaces whose coordinates are labeled as ui and ūi where i = 1, · · · , n.
State the pattern of the row and column indices of the Jacobian matrix in relation to
the indices of the coordinates of the two spaces.
What is the difference between the Jacobian matrix and the Jacobian and what is the
relation between them?
What is the relation between the Jacobian of a given transformation and the Jacobian
of its inverse? Write a mathematical formula representing this relation.
Is the labeling of two coordinate systems (e.g. barred and unbarred) involved in a
transformation relation essential or arbitrary? Hence, discuss if the labeling of the
coordinates in the Jacobian matrix can be interchanged.
Using the transformation equations between the Cartesian and cylindrical coordinate
systems, find the Jacobian matrix of the transformation between these systems, i.e.
Cartesian to cylindrical and cylindrical to Cartesian.
Repeat question 2.32 for the spherical, instead of cylindrical, system to find the Jacobian this time.
Give a simple definition of admissible coordinate transformation.
What is the meaning of the C n continuity condition?
What “invariant” object or property means? Give some illustrating examples.
What is the meaning of “composition of transformations”? State a mathematical
relation representing such a composition.
What is the Jacobian of a composite transformation in terms of the Jacobians of the
simple transformations that make the composite transformation? Write a mathematical relation that links all these Jacobians.
2.8 Exercises and Revision
48
2.39 “The collection of all admissible coordinate transformations with non-vanishing Jacobian form a group”. What this means? State your answer in mathematical and
descriptive forms.
2.40 Is the transformation of coordinates a commutative operation? Justify your answer
by an example.
2.41 A transformation T3 with a Jacobian J3 is a composite transformation, i.e. T3 = T2 T1
where the transformations T1 and T2 have Jacobians J1 and J2 . What is the relation
between J1 , J2 and J3 ?
2.42 Two transformations, R1 and R2 , are related by: R1 R2 = I where I is the identity
transformation. What is the relation between the Jacobians of R1 and R2 ? What we
should call these transformations?
2.43 Discuss the transformation of one set of basis vectors of a given coordinate system to
another set of opposite variance type of that system and the relation of this to the
metric tensor.
2.44 Discuss the transformation of one set of basis vectors of a given coordinate system to
another set of the same variance type of another coordinate system.
2.45 Discuss and compare the results of question 2.43 and question 2.44. Also, compare
the mathematical formulation that should apply in each case.
2.46 Define proper and improper coordinate transformations.
2.47 What is the difference between positive and negative orthogonal transformations?
2.48 Give detailed definitions of coordinate curves and coordinate surfaces of 3D spaces
discussing the relation between them.
2.49 For each one of the following coordinate systems, what is the shape of the coordinate
curves and coordinate surfaces: Cartesian, cylindrical and spherical?
2.50 Make a simple plot representing the φ coordinate curve with the ρ and z coordinate
surfaces of a cylindrical coordinate system.
2.51 Make a simple plot representing the r coordinate curve with the θ and φ coordinate
surfaces of a spherical coordinate system.
2.52 Define “scale factors” of a coordinate system and outline their significance.
2.53 Give the scale factors of the following coordinate systems: orthonormal Cartesian,
cylindrical and spherical.
2.54 Define, mathematically and in words, the covariant and contravariant basis vector
sets explaining any symbols involved in these definitions.
2.55 What is the relation of the covariant and contravariant basis vector sets with the
coordinate curves and coordinate surfaces of a given coordinate system? Make a
simple sketch representing this relation for a general curvilinear coordinate system in
a 3D space.
2.56 The covariant and contravariant components of vectors can be transformed one to the
other. How? State your answer in a mathematical form.
2.57 What is the significance of the following relations?
Ei · Ej = δij
Ei · Ej = δ ij
2.58 Write down the mathematical relations that correlate the basis vectors to the compo-
2.8 Exercises and Revision
49
nents of the metric tensor in their covariant and contravariant forms.
2.59 Using Eq. 50, show that if E1 , E2 , E3 is a right handed orthonormal system then Ei =
Ei . Repeat the question using this time Eq. 51 where E1 , E2 , E3 form a right handed
orthonormal system. Hence, conclude that when the covariant or contravariant basis
vector set is orthonormal then the covariant and contravariant components of a given
tensor are identical.
2.60 State the mathematical relations between the original and transformed (i.e. unbarred
and barred) basis vector sets in their covariant and contravariant forms under admissible coordinate transformations.
2.61 Correct, if necessary, the following equations explaining all the symbols involved:
1
E1 · (E2 × E3 ) = √
g
√
E1 · E2 × E3 = g
2.62 Obtain the relation: g = J 2 from the relation: JT J = [gij ] giving full explanation of
each step.
2.63 State three consequences of having mutually orthogonal contravariant basis vectors
at each point in the space justifying these consequences.
2.64 Discuss the relationship between the concepts of space, coordinate system and metric
tensor.
Chapter 3
Tensors
The subject of this chapter is tensors in general. The chapter is divided into three sections
where we discuss in these sections: tensor types, tensor operations, and tensor representations.
3.1
Tensor Types
In the following subsections, we introduce a number of tensor types and categories and
highlight their main characteristics and differences. These types and categories are not
mutually exclusive and hence they overlap in general. Moreover, they may not be exhaustive in their classes as some tensors may not instantiate any one of a complementary set
of types such as being symmetric or anti-symmetric.
3.1.1
Covariant and Contravariant Tensors
These are the main types of tensor with regard to the rules of their transformation between
different coordinate bases. Covariant tensors are notated with subscript indices (e.g. Ai )
while contravariant tensors are notated with superscript indices (e.g. Aij ). A covariant
tensor is transformed according to the following rule:
∂uj
Aj
∂ ūi
while a contravariant tensor is transformed according to the following rule:
Āi =
(71)
∂ ūi j
A
(72)
∂uj
The barred and unbarred symbols in these equations represent the same mathematical
object (tensor or coordinate) in the transformed and original coordinate systems respectively. An example of covariant tensors is the gradient of a scalar field while an example
of contravariant tensors is the displacement vector.
The above transformation rules, as demonstrated by Eqs. 71 and 72, which correspond
to rank-1 tensors can be easily extended to tensors of any rank following the above pattern.
Hence, a covariant tensor A of rank m is transformed according to the following rule:
Āi =
∂up ∂uq
∂ur
·
·
·
Apq···r
(73)
∂ ūi ∂ ūj
∂ ūm
Similarly, a contravariant tensor B of rank n is transformed according to the following
rule:
∂ ūn pq···r
∂ ūi ∂ ūj
·
·
·
B
(74)
B̄ ij···n =
∂up ∂uq
∂ur
Āij···m =
50
3.1.1 Covariant and Contravariant Tensors
51
Some tensors of rank > 1 have mixed variance type, i.e. they are covariant in some
indices and contravariant in others. In this case the covariant variables are indexed with
subscripts while the contravariant variables are indexed with superscripts, e.g. Aji which
is covariant in i and contravariant in j. A mixed type tensor transforms covariantly in its
covariant indices and contravariantly in its contravariant indices, e.g.
∂ ūl ∂uj ∂ ūn i k
A
(75)
∂ui ∂ ūm ∂uk j
In brief, tensors of rank-n (n > 0) whose all free indices are subscripts are covariant,
and tensors of rank-n (n > 0) whose all free indices are superscripts are contravariant,
while tensors of rank-n (n > 1) whose some of free indices are subscripts while other
free indices are superscripts are mixed. Hence, Ai and Bij are covariant and transform
according to Eq. 73, Ai and B ijk are contravariant and transform according to Eq. 74,
and Aji and Bijk are mixed and transform according to Eq. 75. As indicated before, for
orthonormal Cartesian coordinate systems there is no difference between the covariant and
contravariant representations, and hence the indices can be upper or lower although it is
common to use lower indices in this case.
The practical rules for writing the transformation equations, as seen in the above equations, can be summarized in the following points where we take Eq. 75 (which is sufficiently
general) as an example to illustrate the steps taken in writing the transformation equations:
1. We start by writing the symbol of the transformed tensor on the left hand side of
the transformation equation and the symbol of the original tensor on the right hand
side, that is:
Ā = A
(76)
Ālmn =
2. We index the transformed tensor according to its intended indicial structure observing the order of the indices. We similarly index the original tensor with its original
indices, that is:
Ālmn = Ai jk
(77)
3. We insert a number of partial differential operators on the right hand side equal to
the number of free indices, that is:
∂u ∂u ∂u i k
A
(78)
∂u ∂u ∂u j
4. We index the coordinates of the transformed tensor in the numerator or denominator
in these operators according to the order of the indices in the tensor where these
indices are in the same position (upper or lower) as their position in the tensor, that
is:
∂ul ∂u ∂un i k
A
(79)
Ālmn =
∂u ∂um ∂u j
5. Because the transformed tensor is barred then its coordinates should also be barred,
that is:
∂ ūl ∂u ∂ ūn i k
Ālmn =
A
(80)
∂u ∂ ūm ∂u j
Ālmn =
3.1.1 Covariant and Contravariant Tensors
52
6. We then index the coordinates of the original tensor in the numerator or denominator
in these operators according to the order of the indices in the tensor where these
indices are in the opposite position (upper or lower) to their position in the tensor,
that is:
∂ ūl ∂uj ∂ ūn i k
l n
Ā m =
A
(81)
∂ui ∂ ūm ∂uk j
We note that an upper coordinate index in the denominator of the partial differential operators is equivalent to a lower tensor index. We also note that each pair of the transformed
indices are on the two sides and hence they should be in the same position (upper or lower)
according to the rules of free indices (see § 1.2), while the pairs of indices of the original
tensor are in the same side and hence they should be in opposite positions according to
the rules of dummy indices.
The covariant and contravariant types of a tensor are linked through the metric tensor.
As will be detailed later (refer to § 4.5), the contravariant metric tensor is used for raising
covariant indices of covariant and mixed tensors, e.g.
Ai = g ik Ak
Aij = g ik Akj
(82)
where the contravariant metric tensor g ik is used to raise the covariant index k to become the contravariant index i. Similarly, the covariant metric tensor is used for lowering
contravariant indices of contravariant and mixed tensors, e.g.
Ai = gik Ak
Aij = gik Akj
(83)
where the covariant metric tensor gik is used to lower the contravariant index k to become
the covariant index i. Hence, the metric tensor is used to change the variance type of
tensor indices and for that reason it is labeled as index raising or index lowering or index
shifting operator. We note that in the raising and lowering operations, the metric tensor
acts as an index replacement operator, as well as a shifting operator, by changing the label
of the shifted index, as seen in the above examples where the shifted index k is replaced
with i which is the other index of the metric tensor.
Because it is possible to shift the index position of a tensor by using the covariant and
contravariant types of the metric tensor as an index shifting operator, a given tensor can
be cast into a covariant or a contravariant form, as well as a mixed form in the case of
tensors of rank > 1. However, it should be emphasized that the order of the indices must
be respected in this process, because two tensors with the same indicial structure but with
different indicial order are not equal in general, as stated before. For example:
Aij = gjk Aik 6= gjk Aki = Aj i
(84)
Dots may be inserted to remove any ambiguity about the order of the indices and hence
Ai· j or Aij. means i first and j second while A·ji or Aji. means j first and i second.
A tensor of m free contravariant indices and n free covariant indices may be called type
(m, n) tensor. When one or both variance types are absent, zero is used to refer to the
absent variance type in this notation. Accordingly, Akij is a type (1, 2) tensor, B ik is a
3.1.1 Covariant and Contravariant Tensors
53
ts
type (2, 0) tensor, Cm is a type (0, 1) tensor, and Dpqr
is a type (2, 3) tensor. As we will
see in § 3.1.3, the type may also include the weight as a third entry and hence the type
in this sense is identified by the symbol (m, n, w) where m and n refer to the number of
contravariant and covariant indices respectively while w refers to the weight of the tensor.
It is obvious that the rank of the tensor can be obtained from this notation by adding the
first two entries and hence a tensor of type (m, n) or (m, n, w) is of rank m + n.
As seen earlier, the vectors providing the basis set for a coordinate system are of covariant
type when they are tangent to the coordinate curves, which represent the curves along
which exactly one coordinate varies while the other coordinates are held constant, and
they are of contravariant type when they are perpendicular to the coordinate surfaces
which are the surfaces over which all coordinates vary except one which is held constant.
Formally, the covariant and contravariant basis vectors are given respectively by:
Ei =
∂r
∂ui
Ei = ∇ui
(85)
where r = xj ej is the position vector in Cartesian coordinates and ui is a general coordinate. As indicated earlier, a superscript in the denominator of partial derivatives is
equivalent to a subscript in the numerator and hence the above equation is consistent with
the rules of tensor indices which were outlined in § 1.2.
We also remark that in general the basis vectors, whether covariant or contravariant, are
not necessarily of unit length and/or mutually orthogonal although they may be so. In
fact, there are standard mathematical procedures to orthonormalize the basis set if it is
not and if orthonormalization is needed. For example, the two covariant basis vectors, E1
and E2 , of a 2D space (i.e. surface) can be orthonormalized as follows:
E1 =
E1
E1
=√
|E1 |
g11
E2 =
g11 E2 − g12 E1
√
g11 g
(86)
where g is the determinant of the surface covariant metric tensor, the indexed g are the
coefficients of this tensor, and the underlined vectors are orthonormal basis vectors, that
is:
E2 · E2 = 1
E1 · E2 = 0
(87)
E1 · E1 = 1
This can be verified by conducting the dot products of the last equation using the vectors
defined in Eq. 86.
As seen earlier, the two sets of covariant and contravariant basis vectors are reciprocal
systems and hence they satisfy the following reciprocity relations:
Ei · Ej = δij
Ei · Ej = δ ij
(88)
where the indexed δ is the Kronecker delta tensor (refer to § 4.1) which can be represented
by the unity matrix. The reciprocity of these two sets of basis vectors is illustrated
schematically in Figure 11 for the case of a 2D space.
3.1.1 Covariant and Contravariant Tensors
54
E2
E2
φ
E1
φ
E1
Figure 11: The reciprocity relation between the covariant and contravariant basis vectors
in a 2D space where E1 ⊥ E2 , E1 ⊥ E2 , and |E1 | |E1 | cos φ = |E2 | |E2 | cos φ = 1.
A vector can be represented either by covariant components with contravariant basis
vectors or by contravariant components with covariant basis vectors. For example, a
vector A can be expressed as:
A = Ai Ei
or
A = Ai Ei
(89)
where Ei and Ei are the contravariant and covariant basis vectors respectively. This is
illustrated graphically in Figure 12 for a vector A in a 2D space. The use of the covariant
or contravariant form of the vector representation is a matter of choice and convenience
since these two representations are equivalent as they represent and correctly describe the
same object.
More generally, a tensor of any rank (≥ 1) can be represented covariantly using contravariant basis tensors of that rank, or contravariantly using covariant basis tensors, or
in a mixed form using a mixed basis of opposite type. For example, a rank-2 tensor A can
be written as:
A = Aij Ei Ej = Aij Ei Ej = Aij Ei Ej = Aij Ei Ej
(90)
where Ei Ej , Ei Ej , Ei Ej and Ei Ej are dyadic products of the basis vectors of the presumed
system (refer to § 3.2.3). We remark that dyadic products represent a combination of two
vectors and hence they represent two directions in a certain order. Figure 13 is a graphic
illustration of the nine dyadic products of the three unit basis vectors in a 3D space with
a rectangular Cartesian coordinate system.
Similarly, a rank-3 tensor B can be written as:
B = Bijk Ei Ej Ek = B ijk Ei Ej Ek = Bi jk Ei Ej Ek = B i j k Ei Ej Ek = · · · etc.
(91)
where Ei Ej Ek . . . etc. are triads. More generally, a rank-n tensor C can be written as:
C = Cij···n Ei Ej · · · En = C ij···n Ei Ej · · · En = Ci j···n Ei Ej · · · En = · · · etc.
(92)
3.1.2 True and Pseudo Tensors
55
A2 |E2|
A
E2
A2 E2
E2
E1
A1 E1
E1
A1 |E1|
Figure 12: The representation of a vector A in covariant and contravariant basis vector
sets in a 2D space where the components shown at the four points are with reference to
unit vectors in the given directions, e.g. A1 |E1 | is a component with reference to a unit
vector in the direction of E1 .
where Ei Ej · · · En . . . etc. are n-polyads. We note that the order of the indices in the tensor
bases (i.e. dyads, triads and n-polyads) should be the same as the order of the indices in
the tensor components, as seen in the above examples.
Finally, it should be remarked that the two sets of basis vectors (i.e. covariant and
contravariant), like the components of the tensors themselves, are identical for orthonormal
Cartesian systems. This can be explained by the fact that since the coordinate surfaces
are mutually perpendicular planes then the coordinate curves are mutually perpendicular
straight lines and hence the tangent vectors to the xi coordinate curve is the same as the
perpendicular vector to the xi coordinate surface noting that the coordinates are scaled
uniformly in all directions.
3.1.2
True and Pseudo Tensors
These are also called polar and axial tensors respectively although it is more common
to use these terms for vectors. Pseudo tensors may also be called tensor densities, however the terminology in this part, like many other parts, is not universal. True tensors
are proper or ordinary tensors and hence they are invariant under coordinate transformations, while pseudo tensors are not proper tensors since they do not transform invariantly
as they acquire a minus sign under improper orthogonal transformations which involve
inversion of coordinate axes through the origin of coordinates with a change of system
handedness. Figure 14 demonstrates the behavior of a true vector v and a pseudo vector
p where the former keeps its direction in the space following a reflection of the coordinate
system through the origin of coordinates while the latter reverses its direction following
3.1.2 True and Pseudo Tensors
56
x3
x1
e1 e1
x3
x2
x1
x3
x1
e2 e1
e3 e1
x2
x1
x3
x2
x1
x3
x1
e1 e2
x3
e2 e2
x1
e3 e2
x2
x3
x2
x1
x3
x2
e1 e3
e2 e3
x2
x3
x2
x1
e3 e3
x2
Figure 13: The nine unit dyads associated with the double directions of rank-2 tensors
in a 3D space with a rectangular Cartesian coordinate system. The vectors ei and ej
(i, j = 1, 2, 3) are unit vectors in the directions of coordinate axes where the first indexed
e (blue) represents the first vector of the dyad while the second indexed e (red) represents
the second vector of the dyad. In these nine frames, the first vector is fixed along each
row while the second vector is fixed along each column.
this operation.
Because true and pseudo tensors have different mathematical properties and represent
different types of physical entities, the terms of consistent tensor expressions and equations
should be uniform in their true and pseudo type, i.e. all terms should be true or all terms
should be pseudo. The direct product (refer to § 3.2.3) of true tensors is a true tensor.
The direct product of even number of pseudo tensors is a true tensor, while the direct
product of odd number of pseudo tensors is a pseudo tensor. The direct product of a mix
of true and pseudo tensors is a true or pseudo tensor depending on the number of pseudo
tensors involved in the product as being even or odd respectively.
Similar rules to those of the direct product apply to the cross product, including the curl
operation, involving tensors (which are usually of rank-1) with the addition of a pseudo
3.1.3 Absolute and Relative Tensors
57
x3
v
X2
V
X1
p
O
O
P
x2
x1
X3
Figure 14: The behavior of a true vector (v and V) and a pseudo vector (p and P) on
reflecting the coordinate system in the origin of coordinates. The lower case symbols stand
for the objects in the original system while the upper case symbols stand for the same
objects in the reflected system.
factor for each cross product operation. This factor is contributed by the permutation
tensor which is implicit in the definition of the cross product. As we will see in § 4.2, the
permutation tensor is a pseudo tensor.
In summary, what determines the tensor type (true or pseudo) of the tensor terms
involving direct and cross products is the parity of the multiplicative factors of pseudo type
plus the number of cross product operations involved since each cross product operation
contributes a permutation tensor. We note that inner product (see § 3.2.5) is the result of
a direct product (see § 3.2.3) operation followed by a contraction (see § 3.2.4) and hence
it is like a direct product in this context.
3.1.3
Absolute and Relative Tensors
Considering an arbitrary transformation from a general coordinate system to another, a
tensor of weight w is defined by the following general tensor transformation relation:
Āij...klm...n
∂u
=
∂ ū
w
∂ ūk ∂ud ∂ue
∂uf ab...c
∂ ūi ∂ ūj
·
·
·
·
·
·
A de...f
∂ua ∂ub
∂uc ∂ ūl ∂ ūm
∂ ūn
(93)
where ∂u
symbolizes the Jacobian of the transformation between the two systems (see §
∂ ū
2.3). When w = 0 the tensor is described as an absolute or true tensor, and when w 6= 0
the tensor is described as a relative tensor. When w = −1 the tensor may be described as a
pseudo tensor, while when w = 1 the tensor may be described as a tensor density. We note
that some of these labels are used differently by different authors as the terminology of
tensor calculus is not universally approved and hence the conventions of each author should
3.1.4 Isotropic and Anisotropic Tensors
58
be checked. Also, there is an obvious overlap between this classification (i.e. absolute and
relative) and the previous classification (i.e. true and pseudo) at least according to some
conventions. As indicated earlier (see § 3.1.1), a tensor of m contravariant indices and n
covariant indices may be described as a tensor of type (m, n). This may be extended to
include the weight w as a third entry and hence the type of the tensor is identified by
(m, n, w).
Relative tensors can be added and subtracted (see § 3.2.1) if they have the same indicial
structure and have the same weight; the result is a tensor of the same indicial structure
and weight. Also, relative tensors can be equated if they have the same indicial structure
and weight. In brief, the terms of tensor expressions and equalities should have the same
weight w, whether w is equal to zero or not. Multiplication of relative tensors produces
a relative tensor whose weight is the sum of the weights of the original tensors. Hence, if
the weights are added up to a non-zero value the result is a relative tensor of that weight;
otherwise it is an absolute tensor.
We remark that the statements of the previous paragraph can be generalized by including
w = 0 which corresponds to absolute tensors and hence “relative” in those statements is
more general than being opposite to “absolute”. Accordingly, and from the perspective
of relative tensors (i.e. assuming that other qualifications such as matching in indicial
structure, are met), two absolute tensors can be added or subtracted or equated but an
absolute and a relative tensor (i.e. with w 6= 0) cannot since they are “relative” tensors
with different weights.
3.1.4
Isotropic and Anisotropic Tensors
Isotropic tensors are characterized by the property that the values of their components
are invariant under coordinate transformation by proper rotation of axes. In contrast, the
values of the components of anisotropic tensors are dependent on the orientation of the
coordinate axes. Notable examples of isotropic tensors are scalars (rank-0), the vector
0 (rank-1), Kronecker delta (rank-2) and Levi-Civita tensor (rank ≥ 2). Many tensors
describing physical properties of materials, such as stress and magnetic susceptibility, are
anisotropic.
Direct and inner products (see § 3.2.3 and 3.2.5) of isotropic tensors are isotropic tensors.
The zero tensor of any rank and any dimension is isotropic; therefore if the components
of a tensor vanish in a particular coordinate system they will vanish in all properly and
improperly rotated coordinate systems.[14] Consequently, if the components of two tensors
are identical in a particular coordinate system they are identical in all transformed coordinate systems since the tensor of their difference is a zero tensor and hence it is invariant.
This means that tensor equalities and identities are invariant under coordinate transformations, which is one of the main motivations for the use of tensors in mathematics and
science. As indicated, all rank-0 tensors (scalars) are isotropic. Also, the zero vector, 0,
of any dimension is isotropic; in fact it is the only rank-1 isotropic tensor.
[14]
For improper rotation, this is more general than being isotropic.
3.1.5 Symmetric and Anti-symmetric Tensors
3.1.5
59
Symmetric and Anti-symmetric Tensors
These types of tensor apply to high ranks only (rank ≥ 2) since symmetry and antisymmetry of tensors require in their definition two free indices at least, and hence a
scalar with no index and a vector with a single index do not qualify to be symmetric or
anti-symmetric. Moreover, these types are not exhaustive, even for tensors of rank ≥ 2, as
there are high-rank tensors which are neither symmetric nor anti-symmetric, and hence we
may call them asymmetric tensors although this terminology may include anti-symmetric
tensors as well.
A rank-2 tensor Aij is symmetric in its components iff for all i and j the following
condition is satisfied:
Aji = Aij
(94)
and anti-symmetric or skew-symmetric iff for all i and j the following condition is satisfied:
Aji = −Aij
(95)
Similar conditions apply to contravariant type tensors, i.e. Aji = Aij for symmetric and
Aji = −Aij for anti-symmetric. This also applies to the following definitions and identities
which are largely presented in covariant forms.
More generally, a rank-n tensor Ai1 ...in is symmetric in its two indices ij and il iff the
following condition applies identically:
Ai1 ...il ...ij ...in = Ai1 ...ij ...il ...in
(96)
and anti-symmetric in its two indices ij and il iff the following condition applies identically:
Ai1 ...il ...ij ...in = −Ai1 ...ij ...il ...in
(97)
Any rank-2 tensor Aij can be synthesized from (or decomposed into) a symmetric part
A(ij) , which is marked with round brackets enclosing the indices, and an anti-symmetric
part A[ij] , which is marked with square brackets, where the following relations apply:
Aij = A(ij) + A[ij]
1
(Aij + Aji )
A(ij) =
2
1
(Aij − Aji )
A[ij] =
2
(98)
(99)
(100)
The first relation can be verified by substituting the second and third relations into the
first, that is:
Aij = A(ij) + A[ij] =
1
1
1
(Aij + Aji ) + (Aij − Aji ) = (2Aij ) = Aij
2
2
2
(101)
which is an identity, while the second and third relations can be verified by shifting the
indices, that is:
A(ji) =
1
1
(Aji + Aij ) = (Aij + Aji ) = A(ij)
2
2
(102)
3.1.5 Symmetric and Anti-symmetric Tensors
A[ji] =
1
1
(Aji − Aij ) = − (Aij − Aji ) = −A[ij]
2
2
60
(103)
and hence A(ij) is symmetric and A[ij] is anti-symmetric.
Similarly, a rank-3 tensor Aijk can be symmetrized by the following relation:
A(ijk) =
1
(Aijk + Akij + Ajki + Aikj + Ajik + Akji )
3!
(104)
and anti-symmetrized by the following relation:
A[ijk] =
1
(Aijk + Akij + Ajki − Aikj − Ajik − Akji )
3!
(105)
where the first three terms in these equations are the even permutations of the indices ijk
and the last three terms are the odd permutations of these indices. These relations can
also be verified by exchanging the indices. For example:
A(jik) =
1
(Ajik + Akji + Aikj + Ajki + Aijk + Akij ) = A(ijk)
3!
(106)
and hence it is symmetric in i and j. Similarly:
A[jik] =
1
(Ajik + Akji + Aikj − Ajki − Aijk − Akij ) = −A[ijk]
3!
(107)
and hence it is anti-symmetric in i and j. The exchange of other indices can be done
similarly and hence the symmetric nature of A(ijk) and the anti-symmetric nature of A[ijk]
with respect to their other indices can also be verified. The symmetric and anti-symmetric
parts, A(ijk) and A[jik] , are related by the following equations:
(108)
3 A(ijk) + A[jik] = Aijk + Akij + Ajki
3 A(ijk) − A[jik] = Aikj + Ajik + Akji
(109)
which can also be verified by substitution, as done for rank-2. The right hand side of the
first equation is the sum of the even permutations of the indices ijk while the right hand
side of the second equation is the sum of the odd permutations of these indices.
More generally, a rank-n tensor Ai1 ...in can be symmetrized by:
A(i1 ...in ) =
X
1 X
even permutations of i’s +
odd permutations of i’s
n!
and anti-symmetrized by:
X
1 X
A[i1 ...in ] =
even permutations of i’s −
odd permutations of i’s
n!
(110)
(111)
Similar verifications and relations can be established for the rank-n case following the
patterns seen in the above rank-2 and rank-3 cases.
3.1.5 Symmetric and Anti-symmetric Tensors
61
A tensor of high rank (> 2) may be symmetrized or anti-symmetrized with respect to
only some of its indices instead of all of its indices. For example, in the following the
tensor Aijk is symmetrized and anti-symmetrized respectively with respect to its first two
indices only, that is:
1
(Aijk + Ajik )
2
1
(Aijk − Ajik )
=
2
A(ij)k =
(112)
A[ij]k
(113)
The symmetry of A(ij)k and the anti-symmetry of A[ij]k with respect to the ij indices can
be verified by exchanging these indices in the above relations as done previously for A(ij)
and A[ij] . Moreover, the tensor Aijk can be expressed as the sum of these symmetric and
anti-symmetric tensors, that is:
A(ij)k + A[ij]k =
1
1
(Aijk + Ajik ) + (Aijk − Ajik ) = Aijk
2
2
(114)
A tensor is described as totally symmetric iff it is symmetric with respect to all of its
indices, that is:
Ai1 ...in = A(i1 ...in )
(115)
and hence its anti-symmetric part is zero, i.e. A[i1 ...in ] = 0. Similarly, a tensor is described
as totally anti-symmetric iff it is anti-symmetric in all of its indices, that is:
Ai1 ...in = A[i1 ...in ]
(116)
and hence its symmetric part is zero, i.e. A(i1 ...in ) = 0.
For a totally anti-symmetric tensor, non-zero entries can occur only when all the indices
are different. This is because an exchange of two identical indices, which identifies identical
entries, should change the sign due to the anti-symmetry and no number can be equal to
its negation except zero. Therefore, if the tensor Aijk is totally anti-symmetric then only
its entries of the form Aijk , where i 6= j 6= k, are not identically zero while all the other
entries (i.e. all those of the from Aiij , Aiji , Ajii and Aiii where i 6= j with no sum on i)
vanish identically. In fact, this also applies to partially anti-symmetric tensors where the
entries corresponding to identical anti-symmetric indices should vanish identically. Hence,
if Aijkl is anti-symmetric in its first two indices and in its last two indices then only its
entries with i 6= j and k 6= l do not vanish identically while all its entries with i = j or
k = l (or both) are identically zero.
It should be remarked that the indices whose exchange defines the symmetry and antisymmetry relations should be of the same variance type, i.e. both upper or both lower.
Hence, a tensor Aji is not symmetric if the components satisfy the relation Aji = Aij or
anti-symmetric if the components satisfy the relation Aji = −Aij . This should be obvious
by considering that the covariant and contravariant indices correspond to different basis
sets (i.e. contravariant and covariant), as explained in § 3.1.1.
Another important remark is that the symmetry and anti-symmetry characteristic of a
tensor is invariant under coordinate transformations. Hence, a symmetric/anti-symmetric
3.1.6 General and Affine Tensors
62
tensor in one coordinate system is symmetric/anti-symmetric in all other coordinate systems. This is because if we label the tensor as A and its transformation by symmetry
or anti-symmetry as B, then an equality of the two tensors based on their symmetric or
anti-symmetric nature (i.e. Aij = Bji or Aij = −Bji ) can be expressed, using an algebraic
transformation, as a zero tensor (i.e. Aij − Bji = 0 or Aij + Bji = 0) which is invariant
under coordinate transformations as stated earlier (see § 3.1.4). Similarly, a tensor which
is neither symmetric nor anti-symmetric in one coordinate system should remain so in all
other coordinate systems obtained by permissible transformations. This is based on the
previous statement because if it is symmetric or anti-symmetric in one coordinate system it should be symmetric or anti-symmetric in all other coordinate systems according
to the previous statement. In brief, the characteristic of a tensor as being symmetric or
anti-symmetric or neither is invariant under permissible coordinate transformations.
Finally, for a symmetric tensor Aij and an anti-symmetric tensor B ij (or the other way
around) we have the following useful and widely used identity:
Aij B ij = 0
(117)
This is because an exchange of the i and j indices will change the sign of one tensor
only, which is the anti-symmetric tensor, and this will change the sign of the term in the
summation resulting in having a sum of terms which is identically zero due to the fact
that each term in the sum has its own negation. This also includes the zero components
of the anti-symmetric tensor where the terms containing these components are zero.
3.1.6
General and Affine Tensors
Affine tensors are tensors that correspond to admissible linear coordinate transformations
(see § 2.3.4) from an original rectangular system of coordinates, while general tensors
correspond to any type of admissible coordinate transformations. These categories are
part of the terminology of tensor calculus which we use in this book and hence they do
not have a particular significance.
3.2
Tensor Operations
In this section, we briefly examine the main elementary tensor operations, which are
mostly of algebraic nature, that permeate tensor algebra and calculus. There are various
operations that can be performed on tensors to produce other tensors in general. Examples
of these operations are addition/subtraction, multiplication by a scalar (rank-0 tensor),
multiplication of tensors (each of rank > 0), contraction and permutation. Some of these
operations, such as addition and multiplication, involve more than one tensor while others,
such as permutation, are performed on a single tensor. Contraction can involve one tensor
or two tensors.
Before we start our investigation, we should remark that the last subsection (i.e. §
3.2.7), which is about the quotient rule of tensors, is added to the present section of
tensor operations because this section is the most appropriate place for it in the present
3.2.1 Addition and Subtraction
63
book considering the dependency of the definition of this rule on other tensor operations;
otherwise the subsection is not about a tensor operation in the same sense as the operations
presented in the other subsections. Another remark is that in tensor algebra division is
allowed only for scalars, and hence if the components of an indexed tensor should appear in
a denominator, the tensor should be redefined to avoid this. For example, if Ai represents
the components of a tensor and we should have A1i as the components of another tensor,
then we introduce another tensor, say Bi , whose components are the reciprocals of Ai
(i.e. Bi = A1i ) and use Bi instead of A1i . The purpose of this change of notation is to
avoid confusion and facilitate the manipulation of tensors according to the familiar rules
of indices.
3.2.1
Addition and Subtraction
Tensors of the same rank and type can be added algebraically to produce a tensor of the
same rank and type, e.g.
a=b+c
Ai = B i − C i
Aij = Bji + Cji
(118)
In this operation the entries of the two tensors are added algebraically componentwise.
For example, in the second equality of the last equation the C1 component of the tensor
C is subtracted from the B1 component of the tensor B to produce the A1 component of
the tensor A, while in the third equality the B21 component of the tensor B is added to
the C21 component of the tensor C to produce the A12 component of the tensor A.
We note that “type” in the above statement refers to variance type (covariant, contravariant, mixed) and true/pseudo type as well as other qualifications to which the tensors participating in an addition or subtraction operation should match such as having the same
weight if they are relative tensors, as outlined previously (refer for example to § 3.1.3).
The added/subtracted terms should also have the same indicial structure with regard to
their free indices, as explained in § 1.2 in the context of the rules governing the indices
j
of tensor expressions and equalities. Hence, the tensors Aijk and Bik
cannot be added or
i
subtracted although they are of the same rank and type, but the tensors Aijk and Bjk
can be added and subtracted. In brief, the tensors involved in addition and subtraction
operations should satisfy all the rules that have been stated previously on the terms of
tensor expressions and equalities.
Addition of tensors is associative and commutative, that is:
(A + B) + C = A + (B + C)
A+B = B+A
(119)
(120)
In fact, associativity and commutativity can include subtraction if the minus sign is absorbed in the subtracted tensor; in which case the operation is converted to addition.
3.2.2 Multiplication of Tensor by Scalar
3.2.2
64
Multiplication of Tensor by Scalar
A tensor can be multiplied by a scalar, which should not be zero in general, to produce a
tensor of the same variance type, rank and indicial structure, e.g.
j
Ajik = aBik
(121)
where a is a non-zero scalar. As indicated by this equation, multiplying a tensor by a
scalar means multiplying each component of the tensor by that scalar. Multiplication by
a scalar is commutative and distributive over algebraic sum of tensors, that is:
aA = Aa
a (A ± B) = aA ± aB
(122)
(123)
where a is a scalar and A and B are non-scalar tensors. It is also associative when more
than two factors are involved, that is:
a (bA) = (ab) A
a (A ◦ B) = (aA) ◦ B = A ◦ (aB)
(124)
(125)
where b is another scalar and ◦ represents an outer or inner product operation (see § 3.2.3
and 3.2.5).
3.2.3
Tensor Multiplication
This operation, which can be defined generically as multiplication between two non-scalar
tensors, may also be called outer or exterior or direct or dyadic multiplication, although
some of these names may be reserved to operations on vectors. On multiplying each
component of a tensor of rank r by each component of a tensor of rank k, both of dimension
d, a tensor of rank (r+k) with dr+k components is obtained where the variance type of each
index (covariant or contravariant) is preserved. More clearly, by multiplying a tensor of
type (m1 , n1 , w1 ) by a tensor of type (m2 , n2 , w2 ) a tensor of type (m1 +m2 , n1 +n2 , w1 +w2 )
is obtained. This means that the tensor rank and weight in the outer product operation
are additive and the operation conserves the variance type of each index of the tensors
involved. Also, the order of the indices in the product should match the order of the
indices in the multiplicands, as will be seen in the following examples, since tensor indices
refer to specific tensor bases and the multiplication of tensors is not commutative (refer
to § 3.1.1).
For example, if A and B are covariant tensors of rank-1, then on multiplying A by B
we obtain a covariant tensor C of rank-2 and type (0, 2) where the components of C are
given by:
Cij = Ai Bj
(126)
while on multiplying B by A we obtain a covariant tensor D of rank-2 and type (0, 2)
where the components of D are given by:
Dij = Bi Aj
(127)
3.2.3 Tensor Multiplication
65
Similarly, if A is a contravariant tensor of rank-2 and B is a covariant tensor of rank-2,
then on multiplying A by B we obtain a mixed tensor C of rank-4 and type (2, 2) where
the components of C are given by:
C ijkl = Aij Bkl
(128)
while on multiplying B by A we obtain a mixed tensor D of rank-4 and type (2, 2) where
the components of D are given by:
Dijkl = Bij Akl
(129)
In general, the outer product of tensors yields a tensor. In the outer product operation,
it is generally understood that all the indices of the involved tensors have the same range
(i.e. all the tensors have the same dimension) although this may not always be the case. As
indicated before (see § 1.2), there are cases of tensors which are not uniformly dimensioned,
and in some cases these tensors may be regarded as the result of an outer product of lower
rank tensors.
The direct multiplication of tensors may be marked by the symbol , mostly when using
symbolic notation for tensors, e.g. A B. However, in the present book no symbol will
be used to represent the operation of direct multiplication[15] and hence the operation is
symbolized by putting the symbols of the tensors side by side, e.g. AB where A and B
are non-scalar tensors. In this regard, the reader should be vigilant to avoid confusion
with the operation of matrix multiplication which, according to the notation of matrix
algebra, is also symbolized as AB where A and B are matrices of compatible dimensions,
since matrix multiplication from tensor perspective is an inner product, rather than an
outer product, operation.
The direct multiplication of tensors is not commutative in general as indicated above;
however it is distributive with respect to algebraic sum of tensors, that is:
(130)
AB 6= BA
A (B ± C) = AB ± AC
(B ± C) A = BA ± CA
(131)
Regarding the associativity of direct multiplication, there are cases in which this operation
is not associative according to the research literature of tensor calculus.
As indicated before, the rank-2 tensor constructed by the direct multiplication of two
vectors is commonly called dyad. More generally, tensors may be expressed as an outer
product of vectors where the rank of the resultant product is equal to the number of the
vectors involved, i.e. 2 for dyads, 3 for triads and so on. However, not every tensor can
be synthesized as a product of lower rank tensors. Multiplication of a tensor by a scalar
(refer to § 3.2.2) may be regarded as a special case of direct multiplication since it is a
tensor multiplication operation with one of the tensors involved being of rank-0 although
the definition of direct multiplication seems to restrict this operation to non-scalar tensors,
as stated above.
[15]
We mean “specifically” because we use the symbol ◦ to represent general tensor multiplication which
includes inner and outer tensor products.
3.2.4 Contraction
3.2.4
66
Contraction
The contraction operation of a tensor of rank > 1 is to make two free indices identical,
by unifying their symbols, followed by performing summation over these repeated indices,
e.g.
Aji
contraction
Aii
(132)
−−−−−−−−→
contraction on jl
Amk
Ajk
(133)
im
il
−−−−−−−−−−−−→
Accordingly, contraction results in a reduction of the rank by 2 since it requires the annihilation of two free indices by the summation operation. Therefore, the contraction of a
rank-2 tensor results in a scalar, the contraction of a rank-3 tensor results in a vector, the
contraction of a rank-4 tensor results in a rank-2 tensor, and so on. Contraction can also
take place between two tensors as part of an inner product operation, as will be explained
in § 3.2.5, although the contraction in this case is actually done on the tensor resulting
from the outer product operation that underlies the inner product.
For general coordinate systems, the pair of contracted indices should be different in their
variance type, i.e. one upper and one lower. Hence, contraction of a mixed tensor of type
(m, n) will in general produce a tensor of type (m − 1, n − 1). A tensor of type (p, q) can
therefore have p × q possible contractions, i.e. one contraction for each combination of
lower and upper indices. As indicated before, there is no difference between the covariant
and contravariant types in orthonormal Cartesian systems and hence contraction can take
place between any pair of indices. Accordingly, a rank-n tensor in orthonormal Cartesian
possible individual contraction operations where each one of the n
systems can have n(n−1)
2
indices can be contracted with each one of the remaining (n − 1) indices and the factor 2 in
the denominator represents the fact that the contraction operation is independent of the
order of the two contracted indices since contracting i with j is the same as contracting
j with i.[16] We note that conducting a contraction operation on a tensor results into a
tensor. Similarly, the application of a contraction operation on a relative tensor (see §
3.1.3) produces a relative tensor of the same weight as the original tensor.
A common example of a contraction operation conducted on a single tensor is the operation of taking the trace of a square matrix, as defined in matrix algebra, by summing
its diagonal elements, which can be considered as a contraction operation on the rank-2
tensor represented by the matrix, and hence it yields the trace which is a scalar. Similarly,
a well known example of a contraction operation that takes place between two tensors is
the dot product operation on vectors which can be considered as a direct multiplication
(refer to § 3.2.3) of the two vectors that results in a rank-2 tensor followed by a contraction
operation and hence it produces a scalar.
3.2.5
Inner Product
On taking the outer product (refer to § 3.2.3) of two tensors of rank ≥ 1 followed by a
contraction (refer to § 3.2.4) on two indices of the product, an inner product of the two
[16]
This may also be formulated by the rule of combination.
3.2.5 Inner Product
67
tensors is formed. Hence, if one of the original tensors is of rank-m and the other is of
rank-n, the inner product will be of rank-(m + n − 2). In the symbolic notation of tensor
calculus, the inner product operation is usually symbolized by a single dot between the
two tensors, e.g. A · B, to indicate the contraction operation which follows the outer
multiplication.
In general, the inner product is not commutative. When one or both of the tensors
involved in the inner product are of rank > 1 then the order of the multiplicands does
matter in general, that is:
A · B 6= B · A
(134)
However, the inner product operation is distributive with respect to the algebraic sum of
tensors, that is:
A · (B ± C) = A · B ± A · C
(B ± C) · A = B · A ± C · A
(135)
As indicated before (see § 3.2.4), the dot product of two vectors is an example of the
inner product of tensors, i.e. it is an inner product of two rank-1 tensors to produce a
rank-0 tensor. For example, if a is a covariant vector and b is a contravariant vector, then
their dot product can be depicted as follows:
[ab]ij = ai bj
contraction
−−−−−−−−→
a · b = ai bi
(136)
Another common example, from linear algebra, of inner product is the multiplication of
a matrix representing a rank-2 tensor, by a vector, which is a rank-1 tensor, to produce
a vector. For example, if A is a rank-2 covariant tensor and b is a contravariant vector,
then their inner product can be depicted, according to tensor calculus, as follows:
[Ab]ijk = Aij bk
contraction on jk
−−−−−−−−−−−−−→
[A · b]i = Aij bj
(137)
This operation is equivalent to the aforementioned operation of multiplying a matrix by a
vector as defined in linear algebra. We note that we are using here the symbolic notation
of tensor calculus, rather than the matrix notation, in writing Ab and A · b to represent,
respectively, the outer and inner products. In matrix notation, Ab is used to represent the
product of a matrix by a vector which is an inner product according to the terminology of
tensor calculus. The multiplication of two n × n matrices, as defined in linear algebra, to
produce another n × n matrix is another example of inner product. In this operation, each
one of the matrices involved in the multiplication, as well as the product itself, represents
a rank-2 tensor.
For tensors whose outer product produces a tensor of rank > 2 and type (m, n) where
m, n > 0, various contraction operations between different pairs of indices of opposite
variance type can occur and hence more than one inner product, which are different in
general, can be defined. Moreover, when the outer product produces a tensor of rank
> 3 and type (m, n) where m, n > 1, more than one contraction operation can take place
simultaneously. Based on what we have seen earlier (refer to § 3.2.3 and 3.2.4), the outer
product of a tensor of type (p, q) by a tensor of type (s, t) produces a tensor of type
3.2.6 Permutation
68
(p + s, q + t) and hence (p + s) × (q + t) individual inner product operations can take
place, i.e. one inner product for each combination of lower and upper indices. We note
that for orthonormal Cartesian systems the variance type is irrelevant and hence a rank-n
individual inner product operations which is the number of possible
tensor can have n(n−1)
2
contraction operations, as explained earlier (see § 3.2.4).
There are more specialized types of inner product; some of these may be defined differently by different authors. For example, a double contraction inner product of two rank-2
tensors, A and B, may be defined and denoted by double vertically- or horizontally-aligned
dots (i.e. A : B or A · · B) to indicate double contraction taking place between different
pairs of indices. An instance of these specialized types is the inner product with double
contraction of two dyads which is commonly defined by:
ab : cd = (a · c) (b · d)
(138)
where the single dots in the right hand side of this equation symbolize the conventional
dot product of two vectors. The result of this operation is obviously a scalar since it is
the product of two scalars, as can be seen from the right hand side of the equation.
Some authors may define a different type of double contraction inner product of two
dyads, symbolized by two horizontally-aligned dots, which may be called a transposed
contraction. This type of inner product is given by:
ab · · cd = ab : dc = (a · d) (b · c)
(139)
where the result is also a scalar. The second equality of the last equation is based on Eq.
138. We note that the double inner product operators, i.e. : and ··, are defined by some
authors opposite to the above definitions (i.e. the other way around) and hence one should
be on the lookout for such differences in convention.
For two rank-2 tensors, the aforementioned double contraction inner products are similarly defined as in the case of two dyads, that is:
A : B = Aij B ij
A · · B = Aij B ji
(140)
Inner products with higher multiplicities of contraction can be defined similarly, and hence
they may be regarded as trivial extensions of the inner products with lower contraction
multiplicities. Finally, we note that the inner product of tensors produces a tensor because
the inner product is an outer product operation followed by a contraction operation and
both of these operations on tensors produce tensors, as stated before (see 3.2.3 and 3.2.4).
3.2.6
Permutation
A tensor may be obtained by exchanging the indices of another tensor. For example, Aikj
is a permutation of the tensor Aijk . A common example of the permutation operation of
tensors is the transposition of a matrix representing a rank-2 tensor since the first and
second indices, which represent the rows and columns of the matrix, are exchanged in this
operation. It is obvious that tensor permutation applies only to tensors of rank > 1 since
3.2.7 Tensor Test and Quotient Rule
69
no exchange of indices can take place in a scalar with no index or in a vector with a single
index. Also, permutation may be restricted to indices of the same variance type. The
collection of tensors obtained by permuting the indices of a given tensor may be called
isomers.
3.2.7
Tensor Test and Quotient Rule
Sometimes a tensor-like object may be suspected for being a tensor; in such cases a test
based on what is called the “quotient rule” can be used to clarify the situation. We should
remark that the quotient rule of tensors must not be confused with the quotient rule
of differentiation. According to the quotient rule of tensors, if the inner product of a
suspected tensor by a known tensor is a tensor then the suspect is a tensor. In more
formal terms, if it is not known if A is a tensor or not but it is known that B and C
are tensors; moreover it is known that the following relation holds true in all rotated (i.e.
properly-transformed) coordinate frames:
Apq...k...m Bij...k...n = Cpq...mij...n
(141)
then A is a tensor. Here, A, B and C are respectively of ranks m, n and (m+n−2), where
the rank of C is reduced by 2 due to the contraction on k which can be any index of A
and B independently. We assume, of course, that the rules of contraction of indices, such
as being of opposite variance type in the case of non-Cartesian coordinates, are satisfied in
this operation. The form given by the above equation is based, for simplicity, on assuming
a Cartesian system.
Finally, we should remark that testing a suspected tensor for being a tensor can also
be done by employing the first principles through direct application of the transformation rules to see if the alleged tensor satisfies the transformation rules of tensors or not.
However, using the quotient rule is generally more convenient and requires less work. Another remark is that the quotient rule of tensors may be considered by some authors as a
replacement for the division operation which is not defined for tensors.
3.3
Tensor Representations
So far, we are familiar with the covariant and contravariant (including mixed) representations of tensors. There is still another type of representation, that is the physical
representation which is the common one in the scientific applications of tensor calculus
such as fluid and continuum mechanics. The introduction and employment of the physical
representation of tensors are justified by the fact that the covariant and contravariant
basis vectors, as well as the covariant and contravariant components of a vector, do not
in general have the same physical dimensions as explained earlier (see § 2.2). Moreover,
the basis vectors may not have the same magnitude. This motivates the introduction
of a more standard form of vectors by using physical components (which have the same
dimensions) with normalized basis vectors (which are dimensionless with unit magnitude)
where the metric tensor and the scale factors are employed to facilitate this process. The
3.3 Tensor Representations
70
normalization of the basis vectors is done by dividing each vector by its magnitude. For
example, the normalized covariant and contravariant basis vectors of a general coordinate
system, Êi and Êi , are given by:[17]
Êi =
Ei
|Ei |
Êi =
Ei
|Ei |
(no sum on i)
(142)
which for an orthogonal coordinate system becomes (see Eq. 70):
Ei
Êi = p = hi Ei
g ii
Ei
Ei
Êi = √ =
gii
hi
(no sum on i)
(143)
where gii and g ii are the ith diagonal elements of the covariant and contravariant metric
tensor respectively and hi is the scale factor of the ith coordinate as described previously
(see § 2.5 and 2.6).
Consequently, if the physical components of a contravariant vector are notated with a
hat, then for an orthogonal system we have:
Ei
A = Ai Ei = Âi Êi = Âi √
gii
Âi =
=⇒
√
gii Ai = hi Ai
(no sum) (144)
Ai
Ai
g ii Ai = √ =
gii
hi
(no sum) (145)
Similarly, for the covariant form of the vector we have:
Ei
A = Ai Ei = Âi Êi = Âi p
g ii
=⇒
Âi =
p
These definitions and processes can be easily extended to tensors of higher ranks as we
will see next.
The physical components of higher rank tensors are similarly defined as for rank-1 tensors by considering the basis vectors of the coordinate system of the space where similar
simplifications apply to orthogonal coordinate systems. For example, for a rank-2 tensor
A in an orthogonal coordinate system, the physical components can be represented by:
Âij =
Aij
hi hj
Âij = hi hj Aij
hi Aij
Âij =
hj
(no sum on i or j, with basis Êi Êj )
(146)
(no sum on i or j, with basis Êi Êj )
(147)
(no sum on i or j, with basis Êi Êj )
(148)
On generalizing the above pattern, we conclude that the physical components of a tensor
of type (m, n) in an orthogonal coordinate system are given by:
...am
=
Âab11...b
n
[17]
ha1 . . . ham a1 ...am
A
hb1 . . . hbn b1 ...bn
(no sum on any index)
(149)
We note that the factors in the denominators are scalars and hence the rules of indices are not violated.
3.4 Exercises and Revision
71
where the symbols and basis tensors are defined similarly following the above examples.
As a consequence of the last statements, in a space with a well defined metric any
tensor can be expressed in covariant or contravariant (including mixed) or physical forms
using different sets of basis tensors. Moreover, these forms can be transformed from
each other by using the raising and lowering operators and scale factors. As before, for
orthonormal Cartesian systems the covariant, contravariant and physical components are
the same where the Kronecker delta is the metric tensor. This is because the covariant,
contravariant and physical basis vectors are identical in these systems.
More generally, for orthogonal coordinate systems the two sets of normalized covariant
and contravariant basis vectors are identical as established earlier because the corresponding vectors of each basis set are in the same direction (see § 2.4 and 2.6), and hence the
physical components corresponding to the covariant and contravariant components are
identical as well. Consequently, for orthogonal coordinate systems with orthonormal basis
vectors, the covariant, contravariant and physical components are identical because the
normalized bases corresponding to these three forms are identical.
The physical components of a tensor may be represented by the symbol of the tensor with
subscripts denoting the coordinates of the employed coordinate system. For instance, if A
is a vector in a 3D space with contravariant components Ai or covariant components Ai ,
its physical components in Cartesian, cylindrical, spherical and general curvilinear systems
may be denoted by (Ax , Ay , Az ), (Aρ , Aφ , Az ), (Ar , Aθ , Aφ ) and (Au , Av , Aw ) respectively.
For consistency and dimensional homogeneity, the tensors in scientific applications are
commonly represented by their physical components with a set of normalized basis vectors.
The invariance of the tensor form then guarantees that the same tensor formulation is valid
regardless of any particular coordinate system where standard tensor transformations can
be used to convert from one form to another without affecting the validity and invariance
of the formulation.
3.4
Exercises and Revision
3.1 Define “covariant” and “contravariant” tensors from the perspective of their notation
and their transformation rules.
3.2 Write the transformation relations for covariant and contravariant vectors and for covariant, contravariant and mixed rank-2 tensors between different coordinate systems.
3.3 State the practical rules for writing the transformation relations of tensors between
different coordinate systems.
3.4 What are the raising and lowering operators and how they provide the link between
the covariant and contravariant types?
3.5 A is a tensor of type (m, n) and B is a tensor of type (p, q, w). What this means?
Write these tensors in their indicial form.
3.6 Write the following equations in full tensor notation and explain their significance:
Ei =
∂r
∂ui
Ei = ∇ui
3.4 Exercises and Revision
72
3.7 Write the orthonormalized form of the basis vectors in a 2D general coordinate system.
Verify that these vectors are actually orthonormal.
3.8 Why the following relations are labeled as the reciprocity relations?
Ei · Ej = δij
Ei · Ej = δ ij
3.9 The components of the tensors A, B and C are given by: A.ik. j , B.jn. mq and Ckl. i .
Write these tensors in their full notation that includes their basis tensors.
3.10 A, B and C are tensors of rank-2, rank-3 and rank-4 respectively in a given coordinate
system. Write the components of these tensors with respect to the following basis
tensors: , Ei En , Ei Ek Em and Ej Ei Ek En .
3.11 What “dyad” means? Write all the nine unit dyads associated with the double directions of rank-2 tensors in a 3D space with a rectangular Cartesian coordinate system
(i.e. e1 e1 · · · e3 e3 ).
3.12 Make a simple sketch of the nine dyads of exercise 3.11.
3.13 Compare true and pseudo vectors making a clear distinction between the two with a
simple illustrating plot. Generalize this to tensors of any rank.
3.14 Justify the following statement: “The terms of consistent tensor expressions and equations should be uniform in their true and pseudo type”.
3.15 What is the curl of a pseudo vector from the perspective of true/pseudo qualification?
3.16 Define absolute and relative tensors stating any necessary mathematical relations.
3.17 What is the weight of the product of A and B where A is a tensor of type (1, 2, 2)
and B is a tensor of type (0, 3, −1)?
3.18 Show that the determinant of a rank-2 absolute tensor A is a relative scalar and find
the weight in the case of A being covariant and in the case of A being contravariant.
3.19 Why the tensor terms of tensor expressions and equalities should have the same
weight?
3.20 What “isotropic” and “anisotropic” tensor mean?
3.21 Give an example of an isotropic rank-2 tensor and another example of an anisotropic
rank-3 tensor.
3.22 What is the significance of the fact that the zero tensor of all ranks and all dimensions
is isotropic with regard to the invariance of tensors under coordinate transformations?
3.23 Define “symmetric” and “anti-symmetric” tensor. Why scalars and vectors are not
qualified to be symmetric or anti-symmetric?
3.24 Write the symmetric and anti-symmetric parts of the tensor Aij .
3.25 Write the symmetrization and anti-symmetrization formulae for a rank-n tensor Ai1 ...in .
3.26 Symmetrize and anti- symmetrize the tensor Aijkl with respect to its second and fourth
indices.
3.27 Write the two mathematical conditions for a rank-n tensor Ai1 ...in to be totally symmetric and totally anti-symmetric.
3.28 The tensor Aijk is totally symmetric. How many distinct components it has in a 3D
space?
3.29 The tensor Bijk is totally anti-symmetric. How many identically vanishing components
it has in a 3D space? How many distinct non-identically vanishing components it has
3.4 Exercises and Revision
3.30
3.31
3.32
3.33
3.34
3.35
3.36
3.37
3.38
3.39
3.40
3.41
3.42
3.43
3.44
3.45
3.46
3.47
73
in a 3D space?
Give numeric or symbolic examples of a rank-2 symmetric tensor and a rank-2 skewsymmetric tensor in a 4D space. Count the number of independent non-zero components in each case.
Write the formula for the number of independent components of a rank-2 symmetric
tensor, and the formula for the number of independent non-zero components of a
rank-2 anti-symmetric tensor in nD space.
Explain why the entries corresponding to identical anti-symmetric indices should vanish identically.
Why the indices whose exchange defines the symmetry and anti-symmetry relations
should be of the same variance type?
Discuss the significance of the fact that the symmetry and anti-symmetry characteristic of a tensor is invariant under coordinate transformations and link this to the
invariance of the zero tensor.
Verify the relation Aij B ij = 0, where Aij is a symmetric tensor and B ij is an antisymmetric tensor, by writing the sum in full assuming a 3D space.
Classify the common tensor operations with respect to the number of tensors involved
in these operations.
Which of the following operations are commutative, associative or distributive when
these properties apply: algebraic addition, algebraic subtraction, multiplication by a
scalar, outer multiplication, and inner multiplication?
For question 3.37, write all the required mathematical relations that describe those
properties.
The tensors involved in tensor addition, subtraction or equality should be compatible
in their types. Give all the details about these “types”.
What is the meaning of multiplying a tensor by a scalar in terms of the components
of the tensor?
A tensor of type (m1 , n1 , w1 ) is multiplied by another tensor of type (m2 , n2 , w2 ).
What is the type, the rank and the weight of the product?
We have two tensors: A = Aij Ei Ej and B = B kl Ek El . We also have C = AB and
D = BA. Use the properties of tensor operations to obtain the full expression of C
and D in terms of their components and basis tensors (i.e. C = AB = · · · etc.).
Explain why tensor multiplication, unlike ordinary multiplication of scalars, is not
commutative considering the basis tensors to which the tensors are referred.
The direct product of vectors a and b is ab. Edit the following equation by adding a
simple notation to make it correct without changing the order: ab = ba.
What is the difference in notation between matrix multiplication and tensor multiplication of two tensors, A and B, when we write AB?
Define the contraction operation of tensors. Why this operation cannot be conducted
on scalars and vectors?
In reference to general coordinate systems, a single contraction operation is conducted
on a tensor of type (m, n, w) where m, n > 0. What is the rank, the type and the
weight of the contracted tensor?
3.4 Exercises and Revision
74
3.48 What is the condition that should be satisfied by the two tensor indices involved in a
contraction operation assuming a general coordinate system? What about tensors in
orthonormal Cartesian systems?
3.49 How many individual contraction operations can take place in a tensor of type (m, n, w)
in a general coordinate system? Explain why.
3.50 How many individual contraction operations can take place in a rank-n tensor in an
orthonormal Cartesian coordinate system? Explain why.
3.51 List all the possible single contraction operations that can take place in the tensor
Aijk
lm .
3.52 List all the possible double contraction operations that can take place in the tensor
Aij
kmn .
3.53 Give examples of contraction operation from matrix algebra.
3.54 Show that contracting a rank-n tensor results in a rank-(n − 2) tensor.
3.55 Discuss inner multiplication of tensors as an operation composed of two more simple
operations.
3.56 Give common examples of inner product operation from linear algebra and vector
calculus.
3.57 Why inner product operation is not commutative in general?
3.58 Complete the following equations where A and B are rank-2 tensors of opposite
variance type:
A : B =?
A · · B =?
3.59 Write ab : cd in component form assuming a Cartesian system. Repeat this with
ab · · cd.
3.60 Why the operation of inner multiplication of tensors results in a tensor?
3.61 We have: A = Ai Ei , B = Bj Ej and C = C mn Em En . Find the following tensor
products: AB, AC and BC.
3.62 Referring to the tensors in question 3.61, find the following dot products: B · B, C · A
and C · B.
3.63 Define permutation of tensors giving an example of this operation from matrix algebra.
3.64 State the quotient rule of tensors in words and in a formal mathematical form.
3.65 Why the quotient rule is usually used in tensor tests instead of applying the transformation rules?
3.66 Outline the similarities and differences between the three main forms of tensor representation, i.e. covariant, contravariant and physical.
3.67 Define, mathematically, the physical basis vectors, Êi and Êi , in terms of the covariant
and contravariant basis vectors, Ei and Ei .
hi hj hn ikn
3.68 Correct, if necessary, the following relation: Âikn
jm = hk hm Ajm (no sum on any index)
where A is a tensor in an orthogonal coordinate system.
3.69 Why the normalized covariant, contravariant and physical basis vectors are identical
in orthogonal coordinate systems?
3.70 What is the physical significance of being able to transform one type of tensors to
other types as well as transforming between different coordinate systems?
3.71 Why the physical representation of tensors is usually preferred in the scientific appli-
3.4 Exercises and Revision
75
cations of tensor calculus?
3.72 Give a few common examples of physical representation of tensors in mathematical
and scientific applications.
3.73 What is the advantage of representing the physical components of a tensor (e.g. A)
by the symbol of the tensor with subscripts denoting the coordinates of the employed
coordinate system, e.g. (Ar , Aθ , Aφ ) in spherical coordinate systems?
Chapter 4
Special Tensors
The subject of investigation of this chapter is those tensors that form an essential part of
the tensor calculus theory itself, namely the Kronecker, the permutation and the metric
tensors. The chapter also includes a section devoted to the generalized Kronecker delta
tensor which may be regarded as a bridge between the ordinary Kronecker delta tensor
and the permutation tensor. There is also another section in which a number of important
tensor identities related to the Kronecker or/and permutation tensors are collected and
discussed. Finally, there is a section devoted to some important mathematical definitions
and applications in which special tensors are employed.
The Kronecker and permutation tensors are of particular importance in tensor calculus
due to their distinctive properties and unique transformation attributes. They are numerical tensors with invariant components in all coordinate systems. They enter in the
definition of many mathematical objects in tensor calculus and are used to facilitate the
formulation of many tensor identities. Similarly, the metric tensor is one of the most
important tensors (and may even be the most important) in tensor calculus and its applications, as will be revealed in § 4.5 and other parts of this book. In fact, it permeates
the whole subject of tensor calculus due to its role in characterizing the space and its
involvement in essential mathematical definitions, operations and transformations. For
example, it enters in the definition of many mathematical concepts, such as curve length
and surface area, and facilitates the transformation between covariant and contravariant
types of tensors and their basis vectors.
4.1
Kronecker delta Tensor
As indicated above, there are two types of Kronecker delta tensor: ordinary and generalized. In the present section we discuss the ordinary Kronecker delta tensor and in § 4.4
we investigate the generalized Kronecker delta tensor.
The ordinary Kronecker delta tensor, which may also be called the unit tensor, is a
rank-2 numeric, absolute, symmetric, constant, isotropic tensor in all dimensions. It is
defined in its covariant form as:
(
1
(i = j)
δij =
(i, j = 1, 2, . . . n)
(150)
0
(i 6= j)
where n is the space dimension, and hence it can be considered as the identity tensor or
matrix. For example, in a 3D space the Kronecker delta tensor is given by:
1
0
0
δ11 δ12 δ13
1
0
(151)
[δij ] = δ21 δ22 δ23 = 0
0
0
1
δ31 δ32 δ33
76
4.2 Permutation Tensor
77
The contravariant and mixed forms of the ordinary Kronecker delta tensor, i.e. δ ij and
δij , are similarly defined. Consequently, the numerical values of the components of the
covariant, contravariant and mixed types of the Kronecker delta tensor are the same, that
is:
δij = δ ij = δ ij = δij
(152)
and hence they are all defined by Eq. 150. We note that in the last equation (and in any
similar equation) the numerical values of the components (not the tensors) are equated
and hence this is not a violation of the rules of tensor indices as stated in § 1.2. The tensor
is made of components with reference to a set of basis vectors and hence the equality of
the components of two tensors of different variance type does not imply the equality of
the two tensors since the two basis vector sets to which the two tensors are referred can
be different.
From the above definitions, it can be seen that the Kronecker delta tensor is symmetric,
that is:
δij = δji
δ ij = δ ji
(153)
where i, j = 1, 2, . . . , n. Moreover, the tensor is conserved under all proper and improper
coordinate transformations where “conserved” means that the tensor keeps the numerical
values of its components following a coordinate transformation. Since it is conserved
under proper transformations, it is an isotropic tensor. We note that being conserved
under all transformations is stronger than being isotropic as the former applies even under
improper coordinate transformations while isotropy is restricted, by definition, to proper
transformations (see § 3.1.4). We also used “conserved” rather than “invariant” to indicate
the preservation of the components and to avoid confusion with form-invariance which is
a property that characterizes all tensors.
4.2
Permutation Tensor
The permutation tensor, which is a numeric tensor with constant components, is also
known as the Levi-Civita, anti-symmetric and alternating tensor. We note that the “LeviCivita” label is usually used for the rank-3 permutation tensor. Also some authors distinguish between the permutation tensor and the Levi-Civita tensor even for rank-3. Moreover, some of the common labels and descriptions of the permutation tensor are more
specific to rank-3. Hence, differences in conventions, definitions and labels should be considered when reading the literature of tensor calculus related to the permutation tensor.
The permutation tensor has a rank equal to the number of dimensions of the space.
Hence, a rank-n permutation tensor has nn components. This tensor is characterized by
the following properties:
1. It is numeric tensor and hence the value of its components are: −1, 1 and 0 in all
coordinate systems.
2. The value of any particular component (e.g. ǫ312 ) of this tensor is the same in any
coordinate system and hence it is constant tensor in this sense.
4.2 Permutation Tensor
78
3. It is relative tensor of weight −1 for its covariant form and +1 for its contravariant
form.
4. It is isotropic tensor (see § 3.1.4) since its components are conserved under proper
transformations.
5. It is totally anti-symmetric in each pair of its indices, i.e. it changes sign on swapping
any two of its indices.
6. It is pseudo tensor (see § 3.1.2) since it acquires a minus sign under improper orthogonal transformation of coordinates.
7. The permutation tensor of any rank has only one independent non-vanishing component because all the non-zero components of this tensor are of unity magnitude.
8. The rank-n permutation tensor possesses n! non-zero components which is the number of the non-repetitive permutations of its indices.
9. The rank and the dimension of the permutation tensor are identical and hence in an
nD space it has nn components.
The rank-2 permutation tensor ǫij in its covariant form is defined by:[18]
ǫ12 = 1
ǫ21 = −1
ǫ11 = ǫ22 = 0
Similarly, the rank-3 permutation tensor ǫijk in its covariant form is defined by:
(i, j, k is even permutation of 1,2,3)
1
ǫijk = −1
(i, j, k is odd permutation of 1,2,3)
0
(repeated index)
(154)
(155)
The contravariant form of the rank-2 and rank-3 permutation tensors is similarly defined.
Figure 15 is a graphical illustration of the rank-3 permutation tensor ǫijk while Figure
16, which may be used as a mnemonic device, demonstrates the cyclic nature of the three
even permutations of the indices of the rank-3 permutation tensor and the three odd
permutations of these indices assuming no repetition of indices. The three permutations
in each case are obtained by starting from a given number in the cycle and rotating in the
given sense to obtain the other two numbers in the permutation.
The definition of the rank-n permutation tensor (i.e. ǫi1 i2 ...in and ǫi1 i2 ...in ) is similar to
the definition of the rank-3 permutation tensor with regard to the repetition in its indices
(i1 , i2 , · · · , in ) and being even or odd permutations in their correspondence to (1, 2, · · · , n),
that is:
(i1 , i2 , . . . , in is even permutation of 1, 2, . . . , n)
1
i1 i2 ...in
= −1
ǫi1 i2 ...in = ǫ
(i1 , i2 , . . . , in is odd permutation of 1, 2, . . . , n) (156)
0
(repeated index)
As stated before, equations like this defines the numeric values of the tensor components
and hence they do not violate the rules of indices with regard to their variance type (see
§ 1.2) or the rules of relative tensors with regard to their weight (see § 3.1.3).
[18]
There is no rank-1 permutation tensor as there is no possibility of permutation in a 1D space.
4.2 Permutation Tensor
79
ǫ332
ǫ331
ǫ321
ǫ322
ǫ311
ǫ232
ǫ231
ǫ221
ǫ222
ǫ211
ǫ131
ǫ121
ǫ111
ǫ132
ǫ122
ǫ112
ǫ212
ǫ233
ǫ312
ǫ333
ǫ323
ǫ313
ǫ223
ǫ213
ǫ133
ǫ123
ǫ113
Figure 15: Graphical illustration of the rank-3 permutation tensor ǫijk where circular nodes
represent 0, square nodes represent 1 and triangular nodes represent −1.
As well as the inductive definition of the permutation tensor (as given by Eqs. 154, 155
and 156), the permutation tensor of any rank can also be defined analytically where the
entries of the tensor are calculated from closed form formulae. Accordingly, the values of
the entries of the rank-2 permutation tensor can be calculated from the following closed
form equation:
ǫij = ǫij = (j − i)
(157)
where each one of the indices i, j ranges over 1, 2. Similarly, the numerical values of the
entries of the rank-3 permutation tensor are given by:
ǫijk = ǫijk =
1
(j − i) (k − i) (k − j)
2
(158)
where each one of the indices i, j, k ranges over 1, 2, 3. As for the rank-4 permutation
4.2 Permutation Tensor
3
80
3
1
+1
−1
2
2
Even
Odd
1
Figure 16: Graphical demonstration of the cyclic nature of the even and odd permutations
of the indices of the rank-3 permutation tensor assuming no repetition in indices.
tensor we have:
ǫijkl = ǫijkl =
1
(j − i) (k − i) (l − i) (k − j) (l − j) (l − k)
12
(159)
where each one of the indices i, j, k, l ranges over 1, 2, 3, 4.
More generally, the numerical values of the entries of the rank-n permutation tensor can
be obtained from the following formula:
"
#
n−1
n
Y 1 Y
Y
1
ǫa1 a2 ···an = ǫa1 a2 ···an =
(aj − ai )
(160)
(aj − ai ) =
i!
S(n
−
1)
i=1
j=i+1
1≤i<j≤n
where each one of the indices a1 , · · · , an ranges over 1, · · · , n and S(n − 1) is the super
factorial function of the argument (n − 1) which is defined by:
S(k) =
k
Y
i=1
(161)
i! = 1! · 2! · . . . · k!
A simpler formula for calculating the numerical values of the entries of the rank-n permutation tensor can be obtained from the previous one (Eq. 160) by dropping the magnitude
of the multiplication factors and taking their signs only, that is:
!
Y
Y
sgn (aj − ai ) = sgn
ǫa1 a2 ···an = ǫa1 a2 ···an =
(aj − ai )
(162)
1≤i<j≤n
1≤i<j≤n
where sgn(k) is the sign function of the argument k which is defined by:
(k > 0)
+1
sgn(k) = −1
(k < 0)
0
(k = 0)
(163)
4.2 Permutation Tensor
81
The sign function in Eq. 162 can be expressed in a more direct form for the non-vanishing
entries of the permutation tensor (which correspond to the non-repetitive permutations)
by dividing each argument of the multiplicative factors in Eq. 162 by its absolute value,
noting that none of these factors is zero, and hence Eq. 162 becomes:
ǫa1 a2 ···an = ǫa1 a2 ···an =
(aj − ai )
|aj − ai |
1≤i<j≤n
Y
(aj 6= ai )
(164)
Regarding the vanishing entries, they are easily identified by having repeated indices and
hence all the entries of the tensor are easily evaluated.
As stated above, the permutation tensor is totally anti-symmetric (see § 3.1.5) in each
pair of its indices, i.e. it changes sign on swapping any two of its indices, that is:
ǫi1 ...ik ...il ...in = −ǫi1 ...il ...ik ...in
ǫi1 ...ik ...il ...in = −ǫi1 ...il ...ik ...in
(165)
The reason is that any exchange of two indices requires an even/odd number of singlestep shifts to the right of the first index plus an odd/even number of single-step shifts to
the left of the second index, so the total number of shifts is odd and hence it is an odd
permutation of the original arrangement. This may also be concluded from the closed
form formulae (e.g. Eq. 162) where any exchange will lead to an exchange of the indices
of a single multiplicative factor which leads to sign change.[19] Also, the permutation
tensor is a pseudo tensor since it acquires a minus sign under an improper orthogonal
transformation of coordinates, i.e. inversion of axes that changes the system handedness
(see § 2.3.1). However, it is an isotropic tensor since it is conserved under proper coordinate
transformations.
The permutation tensor may be considered as a contravariant relative tensor of weight
+1 or as a covariant relative tensor of weight −1. Since the contravariant and covariant
types of the permutation tensor are relative tensors, it is desirable to define absolute
covariant and contravariant forms of the permutation tensor. This is done by using the
metric tensor, that is:
ǫi1 ...in =
√
g ǫi1 ...in
1
ǫi1 ...in = √ ǫi1 ...in
g
(166)
where the indexed ǫ and ǫ are respectively the relative and absolute permutation tensors
of the given type, and g is the determinant of the covariant metric tensor gij . Referring
to Eqs. 63 and 93, we see that the ǫ tensors are absolute with w = 0. We note that
the contravariant form of the absolute permutation tensor requires a sign function but the
details are out of the scope of the present text (see Zwillinger in the References). However,
for the rank-3 permutation tensor, which is the one used mostly in the present book, the
above expression stands as it is.
[19]
The non-existence of a mixed type permutation tensor is also related to its totally anti-symmetric
nature.
4.3 Identities Involving Kronecker or/and Permutation Tensors
82
The contravariant and covariant types of the rank-3 permutation tensor are linked,
through the Kronecker delta tensor, by the following relations:
j
i j
ǫijk ǫlmk = δli δm
− δm
δl
(167)
i
δli δm
δni
j
j
δl δm δnj
k
δlk δm
δnk
(168)
ijk
ǫ ǫlmn =
where the indexed δ represent the mixed form of the ordinary Kronecker delta. The last
equation can be generalized to nD spaces as follows:
ǫi1 i2 ···in ǫj1 j2 ···jn =
δji11 δji12 · · · δji1n
δji21 δji22 · · · δji2n
.
.. . .
..
. ..
.
.
δjin1 δjin2 · · · δjinn
(169)
The contravariant and covariant types of the rank-n permutation tensor are also linked by
the following relation:
(170)
ǫi1 i2 ···in ǫi1 i2 ···in = n!
where n is the space dimension. The left hand side of this equation represents the sum of
the products of the corresponding entries of the contravariant and covariant types of the
permutation tensor.
On comparing Eq. 169 and the upcoming Eq. 199 we obtain the following identity which
links the contravariant and covariant types of the permutation tensor to each other and
to the generalized Kronecker delta:
n
ǫi1 ...in ǫj1 ...jn = δji11 ...i
...jn
(171)
More details about these identities will be given in the subsequent sections of this chapter.
4.3
4.3.1
Identities Involving Kronecker or/and Permutation Tensors
Identities Involving Kronecker delta Tensor
When an index of the Kronecker delta tensor is involved in a contraction operation by
repeating an index in another tensor in its own term, the effect of this is to replace the
shared index in the other tensor by the other index of the Kronecker delta, that is:
δij Aj = Ai
δji Aj = Ai
(172)
In such cases the Kronecker delta is described as an index replacement or substitution
operator. Hence, we have:
δij δjk = δik
(173)
4.3.2 Identities Involving Permutation Tensor
Similarly, we have:
δij δjk δki = δik δki = δii = n
83
(174)
where n is the space dimension. The last part of this equation (i.e. δii = n) can be easily
justified by the fact that δii is the trace of the identity tensor considering the summation
convention.
Due to the fact that the coordinates are independent of each other (see § 2.2), we also
have the following identity which is based on the well known rules of partial differentiation:
∂ui
= ∂j ui = ui,j = δji
∂uj
(175)
Hence, in an nD space we obtain the following identity from the identities of Eqs. 175 and
174:
∂i ui = δii = n
(176)
Based on the above identities and facts, the following identity can be shown to apply in
orthonormal Cartesian coordinate systems:
∂xi
∂xj
= δij = δji =
∂xj
∂xi
(177)
This identity is based on the two facts that the coordinates are independent, and the
covariant and contravariant types are the same in orthonormal Cartesian coordinate systems.
Similarly, for a set of orthonormal vectors, such as the basis set of orthonormal Cartesian
system, the following identity can be easily proved:
ei · ej = δij
(178)
where the indexed e represents the orthonormal vectors. This identity is no more than
a symbolic statement of the fact that the vectors in such a set are mutually orthogonal
and each one is of unit length. We note that for orthonormal basis sets the covariant and
contravariant forms are identical as explained earlier.
Finally, the double inner product of two dyads formed by an orthonormal set of vectors
satisfies the following identity (see § 3.2.5):
ei ej : ek el = (ei · ek ) (ej · el ) = δik δjl
(179)
This identity is a combination of Eq. 138 and Eq. 178.
4.3.2
Identities Involving Permutation Tensor
From the definition of the rank-3 permutation tensor, we obtain the following identity
which demonstrates the sense of cyclic order of the non-repetitive permutations of this
tensor:
ǫijk = ǫkij = ǫjki = −ǫikj = −ǫjik = −ǫkji
(180)
4.3.2 Identities Involving Permutation Tensor
84
This identity is also a demonstration of the fact that the rank-3 permutation tensor is
totally anti-symmetric in all of its indices since a shift of any two indices reverses its
sign. Moreover, it reflects the fact that this tensor has only one independent non-zero
component since any one of the non-zero entries, all of which are given by Eq. 180, can be
obtained from any other one of these entries by at most a reversal of sign. This identity
also applies to the contravariant form of the permutation tensor.
We also have the following identity for the rank-n permutation tensor:
ǫi1 i2 ···in ǫi1 i2 ···in = n!
(181)
This identity is based on the fact that the left hand side is the sum of the squares of the
epsilon symbol over all the n! non-repetitive permutations of n different indices where the
value of epsilon of each one of these permutations is either +1 or −1 and hence in both
cases their square is +1. The repetitive permutations are zero and hence they do not
contribute to the sum.
The double inner product of the rank-3 permutation tensor and a symmetric tensor A
is given by the following identity:
ǫijk Aij = ǫijk Aik = ǫijk Ajk = 0
(182)
This is because an exchange of the two indices of A does not affect its value due to the
symmetry of A whereas a similar exchange of these indices in ǫijk results in a sign change;
hence each term in the sum has its own negative and therefore the total sum is identically
zero. This identity also applies to the contravariant permutation tensor with a covariant
symmetric tensor.
Another identity with a trivial outcome that involves the rank-3 permutation tensor and
a vector A is the following:
ǫijk Ai Aj = ǫijk Ai Ak = ǫijk Aj Ak = 0
(183)
This can be explained similarly by the fact that, due to the commutativity of ordinary
multiplication, an exchange of the indices in A’s will not affect the value but a similar
exchange in the corresponding indices of ǫijk will cause a change in sign; hence each term
in the sum has its own negative and therefore the total sum will be zero. This identity
also applies to the contravariant permutation tensor with a covariant vector. We also note
that the identities of Eqs. 182 and 183 similarly apply to other ranks of the permutation
tensor as they are based on the totally anti-symmetric property of this tensor.
Finally, for a set of three orthonormal vectors forming a right handed system, the following identities are satisfied:
ei × ej = ǫijk ek
ei · (ej × ek ) = ǫijk
(184)
(185)
These identities are based, respectively, on the forthcoming definitions of the cross product
(see Eq. 494) and the scalar triple product (see Eq. 495) in tensor notation plus the fact
that these vectors are unit vectors.
4.3.3 Identities Involving Kronecker and Permutation Tensors
85
Table 2: Truth table for the identity of Eq. 186.
i
1
1
1
1
1
1
1
1
2
2
2
2
2
2
2
2
4.3.3
j
1
1
1
1
2
2
2
2
1
1
1
1
2
2
2
2
k
1
1
2
2
1
1
2
2
1
1
2
2
1
1
2
2
l
1
2
1
2
1
2
1
2
1
2
1
2
1
2
1
2
ǫij
0
0
0
0
1
1
1
1
-1
-1
-1
-1
0
0
0
0
ǫkl
0
1
-1
0
0
1
-1
0
0
1
-1
0
0
1
-1
0
ǫij ǫkl
0
0
0
0
0
1
-1
0
0
-1
1
0
0
0
0
0
δki
1
1
0
0
1
1
0
0
0
0
1
1
0
0
1
1
δlj
1
0
1
0
0
1
0
1
1
0
1
0
0
1
0
1
δli
1
0
1
0
1
0
1
0
0
1
0
1
0
1
0
1
δkj
1
1
0
0
0
0
1
1
1
1
0
0
0
0
1
1
δki δlj − δli δkj
0
0
0
0
0
1
-1
0
0
-1
1
0
0
0
0
0
Identities Involving Kronecker and Permutation Tensors
For the rank-2 permutation tensor, we have the following identity which involves the
ordinary Kronecker delta tensor in 2D:
ǫij ǫkl =
δki δli
δkj δlj
= δki δlj − δli δkj
(186)
This identity can be proved inductively by building a table for the values on the left
and right hand sides as the indices are varied, as seen in Table 2. The pattern of the
indices in the determinant array of this identity is simple, that is the indices of the first
ǫ provide the indices for the rows as the upper indices of the deltas while the indices of
the second ǫ provide the indices for the columns as the lower indices of the deltas. In
fact, the role of these indices in indexing the rows and columns can be shifted. This can
be explained by the fact that the positions of the two epsilons can be exchanged, since
ordinary multiplication is commutative, and hence the role of the epsilons in providing
the indices for the rows and columns will be shifted. This can also be done by taking
the transposition of the array of the determinant,
which does not change the value of the
T
determinant since det (A) = det A .
We note that a table like Table 2 is similar to the truth tables used in verifying Boolean
and logic identities, and for this reason we label it as a “truth table”. We also note that the
method of building a “truth table” like Table 2 can also be used for proving other similar
identities. The main advantage of this method is that it is easy and straightforward while
its main disadvantage is that it is lengthy and hence it may not be feasible for some messy
4.3.3 Identities Involving Kronecker and Permutation Tensors
86
identities. Moreover, it can provide proofs for special cases but it is not general with
respect to proving similar identities in a general nD space where n is variable although it
may be used as a part of an inductive proof. Another advantage of this method is that it
lends itself to programming and hence it is ideal for use in computing. We note that the
subject matter in these tables is the tensor components.
Another useful identity involving the rank-2 permutation tensor with the Kronecker delta
tensor in 2D is the following:
ǫil ǫkl = δki
(187)
This can be obtained from the identity of Eq. 186 by replacing j with l followed by
minimal algebraic manipulations using tensor calculus rules, that is:
ǫil ǫkl = δki δll − δli δkl
(Eq. 186 with j → l)
= 2δki − δli δkl
=
=
2δki
δki
−
(188)
(Eq. 174)
δki
(Eq. 173)
Similarly, we have the following identity which correlates the rank-3 permutation tensor
to the Kronecker delta tensor in 3D:
ijk
ǫ ǫlmn =
i
δli δm
δni
j
j
δl δm δnj
k
δlk δm
δnk
=
k
j k
i j k
δn + δm
δn δl + δni δlj δm
δli δm
−
j k
k
i j k
δl
δli δnj δm
+ δm
δl δn + δni δm
(189)
Again, the indices in the determinant of this identity follow the same pattern as that of
Eq. 186. Another pattern can also be seen in the six terms on the right where the three
upper indices of all terms are ijk while the three lower indices in the positive terms are
the even permutations of lmn and the three lower indices in the negative terms are the
odd permutations of lmn. This identity can also be established by a truth table similar
to Table 2.
More generally, the determinantal form of Eqs. 186 and 189, which link the rank-2 and
rank-3 permutation tensors to the ordinary Kronecker delta tensors in 2D and 3D spaces,
can be extended to link the rank-n permutation tensor to the ordinary Kronecker delta
tensor in an nD space, that is:
ǫi1 i2 ···in ǫj1 j2 ···jn =
δji11 δji12 · · · δji1n
δji21 δji22 · · · δji2n
..
.. . .
.
. ..
.
.
δjin1 δjin2 · · · δjinn
(190)
Again, the pattern of the indices in the determinant of this identity in their relation to the
indices of the two epsilons follow the same rules as those of Eqs. 186 and 189. Moreover,
the proofs of Eqs. 186 and 189 may be extended to Eq. 190 by induction.
4.3.3 Identities Involving Kronecker and Permutation Tensors
87
Another useful identity in this category is the following:
ǫijk ǫlmk =
i
δli δm
j
j
δl δm
j
i j
= δli δm
− δm
δl
(191)
This identity can be obtained from the identity of Eq. 189 by replacing n with k, that is:
j k
k
k
i j k
i j k
j k
ǫijk ǫlmk = δli δm
− δli δkj δm
δk + δm
δk δl + δki δlj δm
− δm
δl δk − δki δm
δl
j j
i j
i j
i j
i j
i j
= 3δl δm + δm δl + δm δl − δl δm − 3δm δl − δl δm
j
i j
= δli δm
− δm
δl
i
i
δl δm
=
j
δlj δm
(192)
The pattern of the indices in this identity (Eq. 191) is as before if we exclude the repetitive
index k.
The identity of Eq. 191, which may be called the epsilon-delta identity or the contracted
epsilon identity or the Levi-Civita identity, is very useful in manipulating and simplifying
tensor expressions and proving vector and tensor identities; examples of which will be seen
in § 7.1.5. We note that the determinantal form, seen in the middle equality of Eq. 191,
can be considered as a mnemonic device for this identity where the first and second indices
of the first ǫ index the rows while the first and second indices of the second ǫ index the
columns, as given above. In fact, the determinantal form in all the above equations which
are given in this form is a mnemonic device for all these equations, and not only Eq. 191,
where the expanded form, if needed, can be easily obtained from the determinant which
can be easily built following the simple pattern of indices, as explained above.
Other common identities in this category are:
ǫijk ǫljk = 2δli
ǫijk ǫijk = 2δii = 2 × 3 = 3! = 6
(193)
(194)
The first of these identities can be obtained from Eq. 191 with the replacement of m with
j followed by some basic tensor manipulation, that is:
ǫijk ǫljk = δli δjj − δji δlj
= 3δli − δli
=
(Eq. 191 with m → j)
(195)
(Eqs. 174 and 173)
2δli
while the second of these identities can be obtained from the first by replacing l with i
and applying the summation convention in 3D on the right hand side, i.e. using Eq. 174.
The second identity is, in fact, an instance of Eq. 181 for a 3D space.
Another common identity that involves the rank-3 permutation tensor with the ordinary
Kronecker delta in a 3D space is the following:
ǫijk δ1i δ2j δ3k = ǫ123 = 1
(196)
4.4 Generalized Kronecker delta Tensor
88
This identity is based in its first part on the use of the ordinary Kronecker delta as an
index replacement operator (Eq. 172), where each one of the deltas replaces an index in
the permutation tensor, and is based in its second part on the definition of the permutation
tensor (Eq. 155).
Finally, the following identity can also be obtained from the definition of the rank-3
permutation tensor (Eq. 155) and the use of the ordinary Kronecker delta as an index
replacement operator (Eq. 172):
ǫijk δji = ǫijk δki = ǫijk δij = ǫijk δkj = ǫijk δik = ǫijk δjk = 0
(197)
This is because on replacing one of the indices of the permutation tensor it will have two
identical indices and hence it is zero, e.g. ǫijk δji = ǫjjk = 0. We note that identities like
Eqs. 196 and 197 also apply to the opposite variance type. Also, Eqs. 196 and 197 apply
to the permutation tensor of other ranks with some simple modifications.
4.4
Generalized Kronecker delta Tensor
The generalized Kronecker delta tensor in an nD space is an absolute rank-2n tensor of
type (n, n) which is normally defined inductively by:
(j1 . . . jn is even permutation of i1 . . . in )
1
i1 ...in
δj1 ...jn = −1
(198)
(j1 . . . jn is odd permutation of i1 . . . in )
0
(repeated i’s or j’s)
It can also be defined analytically by the following n × n determinant:
n
δji11 ...i
...jn =
δji11 δji12 · · · δji1n
δji21 δji22 · · · δji2n
..
.. . .
.
. ..
.
.
δjin1 δjin2 · · · δjinn
(199)
where the δji entries in the determinant are the ordinary Kronecker deltas as defined
previously (see § 4.1). In this equation, the pattern of the indices in the generalized
n
Kronecker delta symbol δji11 ...i
...jn in connection to the indices in the determinant is similar to
n
the previous patterns seen in § 4.3.3, that is the upper indices in δji11 ...i
...jn provide the upper
indices in the ordinary deltas by indexing the rows of the determinant, while the lower
n
indices in δji11 ...i
...jn provide the lower indices in the ordinary deltas by indexing the columns
of the determinant.
From the previous identities, it can be shown that:
ǫi1 ...in ǫj1 ...jn =
δji11 δji12 · · · δji1n
δji21 δji22 · · · δji2n
..
.. . .
.
. ..
.
.
δjin1 δjin2 · · · δjinn
(200)
4.4 Generalized Kronecker delta Tensor
89
Now, on comparing the last equation with the definition of the generalized Kronecker
delta, i.e. Eq. 199, we conclude that:
n
ǫi1 ...in ǫj1 ...jn = δji11 ...i
...jn
(201)
Based on Eq. 201, the generalized Kronecker delta is the result of multiplying two relative
tensors one of weight w = +1 and the other of weight w = −1 and hence the generalized
Kronecker delta has a weight of w = 0. This shows that the generalized Kronecker delta is
an absolute tensor, as stated above. We remark that the multiplication of relative tensors
produces a tensor whose weight is the sum of the weights of the original tensors (see §
3.1.3).
From Eq. 201, we can see that the relation between the rank-n permutation tensor in
its covariant and contravariant forms and the generalized Kronecker delta in an nD space
is given by:
n
ǫi1 ...in = ǫ1 ... n ǫi1 ...in = δi11...
...in
(202)
...in
ǫi1 ...in = ǫi1 ...in ǫ1 ... n = δ i11...
n
(203)
where the first of these equations is obtained from Eq. 201 by substituting (1 . . . n) for
(i1 . . . in ) in the two sides with relabeling j as i and noting that ǫ1 ... n = 1, while the second
equation is obtained from Eq. 201 by substituting (1 . . . n) for (j1 . . . jn ) in the two sides
and noting that ǫ1 ... n = 1.
Based on Eqs. 202 and 203, the permutation tensor can be considered as an instance
of the generalized Kronecker delta. Consequently, the rank-n permutation tensor can be
written as an n × n determinant consisting of the ordinary Kronecker deltas (Eq. 199).
Moreover, Eqs. 202 and 203 can provide another definition for the permutation tensor in
its covariant and contravariant forms, in addition to the previous inductive and analytic
definitions of this tensor which are given by Eqs. 156 and 160.
ijk
. If we replace n with
The 3D generalized Kronecker delta may be symbolized by δlmn
k and use the determinantal definition of the generalized Kronecker delta of Eq. 199
followed by conducting a few basic algebraic manipulations using some of the above tensor
identities, we obtain:
ijk
=
δlmk
i
δli δm
δki
j
δlj δm
δkj
k
δlk δm
δkk
(Eq. 199)
k
j k
i
j k
k
− δm
δl
δkj δlk − δlj δkk + δki δlj δm
+ δm
δk − δkj δm
= δli δm
j k
k
i j k
i j k
k
j k
= δli δm
δk − δli δkj δm
+ δm
δk δl − δm
δl δk + δki δlj δm
− δki δm
δl
j k
j
i j
i j k
i j
j
= δli δm
δk − δli δm
+ δm
δl − δm
δl δk + δm
δl − δli δm
j
i j
i j
j
j
i j
δl − δli δm
δl + δm
= 3δli δm
− δli δm
+ δm
δl − 3δm
j
= δli δm
=
δli
δlj
i j
− δm
δl
i
δm
j
δm
(Eq. 173)
(Eq. 174)
(204)
4.4 Generalized Kronecker delta Tensor
90
ij
= δlm
(Eq. 199)
that is:
ij
ijk
= δlm
δlmk
(205)
Similar identities can be obtained from contracting two corresponding indices of the nD
generalized Kronecker delta to obtain (n − 1)D generalized Kronecker delta, e.g.
δljij = δli
(206)
which can be simply verified following the former example, that is:
δljij
=
δli δji
δlj δjj
= δli δjj − δji δlj = 2δli − δli = δli
(207)
Returning to the widely used epsilon-delta identity of Eq. 191, if we consider Eq. 205
plus the above identities which correlate the permutation tensor, the generalized Kronecker
delta tensor and the ordinary Kronecker delta tensor, then an identity equivalent to Eq.
191 that involves only the generalized and ordinary Kronecker deltas can be obtained, that
is:[20]
ijk
ǫijk ǫlmk = δlmk
=
=
=
ij
δlm
δli
δlj
j
δli δm
(Eq. 201)
(208)
(Eq. 205)
i
δm
j
δm
(Eq. 199)
i j
− δm
δl
This means that the following relation:
ijk
ij
j
i j
= δlm
= δli δm
− δm
δl
δlmk
(209)
which is no more than the definition of the generalized Kronecker delta of Eq. 199 is
another form of the epsilon-delta identity. The pattern of the indices on the right hand
side in relation to the indices of the generalized Kronecker delta is very simple, that is
we take first the corresponding upper and lower indices followed by the diagonally crossed
indices (i.e. | | − ×). It is worth noting that the epsilon-delta identity (Eqs. 191 and 209)
can also be expressed in a more general form by employing the metric tensor with the
absolute permutation tensor, that is:
g ij ǫikl ǫjmn = gkm gln − gkn glm
(210)
Other identities involving the permutation tensor and the ordinary Kronecker delta tensor can also be formulated in terms of the generalized Kronecker delta tensor.
[20]
In fact, this is a derivation of Eq. 191.
4.5 Metric Tensor
4.5
91
Metric Tensor
The metric tensor, which may also be called the fundamental tensor, is a rank-2 symmetric absolute non-singular tensor, where “non-singular” means invertible and hence its
determinant does not vanish at any point in the space. The metric tensor is one of the
most important special tensors in tensor calculus, if not the most important of all. Its
versatile usage and functionalities permeate the whole discipline of tensor calculus and
its applications. One of the main objectives of the metric is to generalize the concept
of distance to general coordinate systems and hence maintain the invariance of distance
in different coordinate systems, as will be explained next. This tensor is also used to
raise and lower indices and thus facilitate the transformation between the covariant and
contravariant types. As a tensor, the metric has significance regardless of any coordinate
system although it requires a coordinate system to be represented in a specific form (see §
2.7). So, in principle the coordinate system and the space metric are independent entities.
In an orthonormal Cartesian coordinate system of an nD space the length of infinitesimal
element of arc, ds, connecting two neighboring points in space, one with coordinates xi
and the other with coordinates xi + dxi (i = 1, · · · , n), is given by:
(ds)2 = dxi dxi = δij dxi dxj
(211)
This definition of distance is the key to introducing a rank-2 tensor, gij , called the metric
tensor which, for a general coordinate system, is defined by:
(ds)2 = gij dui duj
(212)
where the indexed u represents general coordinates. The metric tensor in the last equation
is of covariant form. The metric tensor has also a contravariant form which is notated
with g ij . It is common to reserve the term “metric tensor” to the covariant form and call
the contravariant form, which is its inverse, the associate or conjugate or reciprocal metric
tensor.
The components of the metric tensor in its covariant and contravariant forms are closely
related to the basis vectors, that is:
gij = Ei · Ej
g ij = Ei · Ej
(213)
(214)
where the indexed E are the covariant and contravariant basis vectors as defined previously
(see § 2.6). Because of these relations, the vectors Ei and Ei may be denoted by gi and
gi respectively which is more suggestive of their relation to the metric tensor. Similarly,
the mixed type metric tensor is given by:
g ij = Ei · Ej = δ ij
g i j = Ei · Ej = δ i j
(215)
and hence it is the identity tensor. These equalities, which may be described as the
reciprocity relations, represent the fact that the covariant and contravariant basis vectors
are reciprocal sets.
4.5 Metric Tensor
92
As a consequence, the covariant metric tensor is given, in full tensor notation, by:
gij =
∂xk ∂xk
∂ui ∂uj
(216)
where
xk = xk u1 , . . . , u n
(k = 1, . . . , n)
(217)
are independent coordinates in an nD space with a rectangular Cartesian system, and ui
and uj (i, j = 1, . . . , n) are independent general coordinates. Similarly, for the contravariant metric tensor we have:
∂ui ∂uj
g ij =
(218)
∂xk ∂xk
As stated already, the basis vectors, whether covariant or contravariant, in general coordinate systems are not necessarily mutually orthogonal and hence the metric tensor is
not diagonal in general since the dot products given by Eqs. 213 and 214 (or Eqs. 216
and 218) are not necessarily zero when i 6= j. Moreover, since those basis vectors vary in
general in their magnitude and relative orientations and they are not necessarily of unit
length, the entries of the metric tensor, including the diagonal elements, are not necessarily of unit magnitude. Also, the entries of the metric tensor, including the diagonal
elements, can be positive or negative.[21] However, since the dot product of vectors is
a commutative operation, the metric tensor is necessarily symmetric. We note that the
mixed type metric tensor is diagonal (or in fact the unity tensor) because the covariant
and contravariant basis vector sets are reciprocal systems (see Eq. 215). This applies in
all coordinate systems.
As indicated above, the covariant and contravariant forms of the metric tensor are inverses of each other and hence we have the following relations:
ij
−1
(219)
g = [gij ]−1
[gij ] = g ij
Hence:
g ik gkj = δ ij
gik g kj = δi j
(220)
where these equations can be seen to represent matrix multiplication (row×column). A
result that can be obtained from the previous statements plus Eqs. 138, 213 and 214 is
that:
(221)
Ei Ej : Ej Ek = Ei · Ej (Ej · Ek ) = g ij gjk = δ ik
jk
k
j
k
j
k
= gij g = δi
(222)
Ei E : Ej E = (Ei · Ej ) E · E
Since the metric tensor has an inverse, it should be non-singular and hence its determinant,
which in general is a function of coordinates like the metric tensor itself, should not vanish
at any point in the space, that is:
g(u1 , . . . , un ) = det (gij ) 6= 0
[21]
(223)
The diagonal entries can be negative when the coordinates are imaginary (see § 2.2.3 and Eq. 241).
4.5 Metric Tensor
93
From the previous statements, it may be concluded that the metric tensor is in fact
a transformation of the Kronecker delta tensor in its different variance types from an
orthonormal Cartesian coordinate system to a general coordinate system, that is (see Eqs.
73-75):
∂xk ∂xl
δkl =
∂ui ∂uj
∂ui ∂uj kl
δ =
g ij =
∂xk ∂xl
∂ui ∂xl k
gji =
δ =
∂xk ∂uj l
gij =
∂xk ∂xk
= Ei · Ej
∂ui ∂uj
∂ui ∂uj
= Ei · Ej
k
k
∂x ∂x
∂ui ∂xk
= Ei · Ej
∂xk ∂uj
(covariant)
(224)
(contravariant)
(225)
(mixed)
(226)
As stated above, the metric tensor is symmetric in its two indices, that is:
g ij = g ji
gij = gji
(227)
This can be easily explained by the commutativity of the dot product of vectors in reference
to the above equations involving the dot product of the basis vectors (Eqs. 213 and 214).
Because of the relations:
Ai = A · Ei = Aj Ej · Ei = Aj g ji
Ai = A · Ei = Aj Ej · Ei = Aj gji
(228)
(229)
the metric tensor is used as an operator for raising and lowering indices and hence facilitating the transformation between the covariant and contravariant types of vectors. By a
similar argument, the above can be generalized where the contravariant metric tensor is
used for raising covariant indices of covariant and mixed tensors and the covariant metric
tensor is used for lowering contravariant indices of contravariant and mixed tensors of any
rank, e.g.
Aik = g ij Ajk
Aikl = gij Ajkl
(230)
Consequently, any tensor in a Riemannian space with well-defined metric can be cast into
covariant or contravariant or mixed forms where for the mixed form the rank should be
> 1. We note that in the operations of raising and lowering of indices the metric tensor
acts, like the Kronecker delta tensor, as an index replacement operator beside its action
in shifting the index position.
In this context, it should be emphasized that the order of the raised and lowered indices
is important and hence:
g ik Ajk = Aj i
and
g ik Akj = Ai j
(231)
are different in general. A dot may be used to indicate the original position of the shifted
index and hence the order of the indices is recorded, e.g. Aj i· and Ai· j for the above
examples respectively. Because raising and lowering of indices is a reversible process,
keeping a record of the original position of the shifted indices will facilitate the reversal if
4.5 Metric Tensor
94
needed. We note that dots may also be inserted in the symbols of mixed tensors to remove
any ambiguity about the order of the indices even without the action of the raising and
lowering operators (refer to § 1.2).
For a space with a coordinate system in which the metric tensor can be cast into a
diagonal form with all the diagonal entries being of unity magnitude (i.e. ±1) the space
and the metric are called flat. For example, in 2D manifolds a plane surface is a flat
space since it can be coordinated by an orthonormal 2D Cartesian system which results
into a diagonal unity metric tensor since the basis vectors of this system are mutually
perpendicular constant vectors and each is of unit length. On the other hand, an ellipsoid
is not a flat space (i.e. it is curved) because due to its intrinsic curvature it is not possible
to find a coordinate system for this type of surface whose basis vectors produce a diagonal
metric tensor with constant diagonal elements of unity magnitude. In this context we note
that a Riemannian metric, gij , in a particular coordinate system is a Euclidean metric if it
can be transformed to the identity tensor, δij , by a permissible coordinate transformation.
If g and ḡ are the determinants of the covariant metric tensor in unbarred and barred
coordinate systems respectively, i.e. g = det (gij ) and ḡ = det (ḡij ), then we have:
√
√
ḡ = J 2 g
ḡ = J g
(232)
where J (= ∂u
) is the Jacobian of the transformation between the unbarred and barred
∂ ū
systems. Consequently, the determinant of the covariant metric tensor and its square root
are relative scalars of weight +2 and +1 respectively (see § 3.1.3).
A “conjugate” or “associate” tensor of a tensor in a metric space is a tensor obtained
by inner product multiplication, once or more, of the original tensor by the covariant or
contravariant forms of the metric tensor. All tensors associated with a particular tensor
through the metric tensor represent the same tensor but in different base systems since
the association is no more than raising or lowering indices by the metric tensor which is
equivalent to a representation of the components of the tensor relative to different basis
sets.
A sufficient and necessary condition for the components of the metric tensor to be constants in a given coordinate system is that the Christoffel symbols of the first or second
kind vanish identically (refer to 5.1). This may be concluded from Eqs. 307 and 308. The
metric tensor behaves as a constant with respect to covariant and absolute differentiation
(see § 5.2 and § 5.3). Hence, in all coordinate systems the covariant and absolute derivatives of the metric tensor are zero. Accordingly, the covariant and absolute derivative
operators bypass the metric tensor in differentiating inner and outer products of tensors
involving the metric tensor.
In orthogonal coordinate systems in an nD space the metric tensor in its covariant and
contravariant forms is diagonal with non-vanishing diagonal elements, that is:
gij = 0
gii 6= 0
g ij = 0
ii
g 6= 0
(i 6= j)
(no sum on i)
(233)
(234)
4.5 Metric Tensor
95
Moreover, we have:
gii = (hi )2 =
1
g ii
det (gij ) = g = g11 g22 . . . gnn =
(no sum on i)
Y
(hi )2
(235)
(236)
i
det g
ij
"
Y
1
1
=
(hi )2
= =
g
g11 g22 . . . gnn
i
#−1
(237)
where hi (= |Ei |) are the scale factors, as described previously in § 2.5 and 2.6.
As indicated before, for orthonormal Cartesian coordinate systems in a 3D space, the
metric tensor is given in its covariant and contravariant forms by the 3 × 3 unit matrix,
that is:
1
0
0
1
0 = δ ij = g ij
(238)
[gij ] = [δij ] = 0
0
0
1
For cylindrical coordinate systems of 3D spaces identified by the coordinates (ρ, φ, z), the
metric tensor is given in its covariant and contravariant forms by:
1
0
0
1
0
0
1
0
ρ2
0
g ij = 0
(239)
[gij ] = 0
ρ2
0
0
1
0
0
1
while for spherical coordinate systems of 3D spaces identified by the coordinates (r, θ, φ),
the metric tensor is given in its covariant and contravariant forms by:
1
0
0
1 0
0
ij
0
0
g = 0 r12
[gij ] = 0 r2
(240)
2
1
2
0
0 r2 sin2 θ
0 0 r sin θ
As seen in Eqs. 238-240, all these metric tensors are diagonal since all these coordinate
systems are orthogonal. We also notice that all the corresponding diagonal elements of
the covariant and contravariant types are reciprocals of each other. This can be easily
explained by the fact that these two types are inverses of each other, moreover the inverse
of an invertible diagonal matrix is a diagonal matrix obtained by taking the reciprocal
of the corresponding diagonal elements of the original matrix, as it is known from linear
algebra. We also see that the diagonal elements are the squares of the scale factors of
these systems or their reciprocals (refer to Table 1).
The Minkowski metric, which is the metric tensor of the 4D space-time of the mechanics
of Lorentz transformations, is given by one of the following two forms:
1
0
0
0
−1
0
0
0
0 −1
ij 0
0
0
1
0
0
(241)
=
[gij ] = g ij =
[g
]
=
g
ij
0
0
0 −1
0
0
1
0
0
0
0 −1
0
0
0
1
4.5 Metric Tensor
96
Consequently, the length of line element ds can be imaginary (see Eqs. 36 and 212). As
seen, this metric belongs to a flat space since it is diagonal with all the diagonal entries
being ±1.
The partial derivatives of the components of the covariant and contravariant metric
tensor satisfy the following identities:
∂k gij = −gmj gni ∂k g nm
∂k g ij = −g mj g in ∂k gnm
(242)
(243)
A related formula for the partial derivatives of the components of the covariant and contravariant metric tensor is given by:
gim ∂k g mj = −g mj ∂k gim
(244)
This relation can be obtained by partial differentiation of the relation gim g mj = δij (Eq.
220) with respect to the k th coordinate using the product rule, that is:
=⇒
gim ∂k g mj = −g mj ∂k gim (245)
∂k gim g mj = gim ∂k g mj + g mj ∂k gim = ∂k δij = 0
The last step in the differentiation (i.e. ∂k δij = 0) is justified by the fact that the components of the Kronecker delta tensor are constants.
In fact, Eq. 242 can be obtained form Eq. 244 by relabeling m as n and multiplying
both sides of Eq. 244 with gmj followed by contraction and exchanging the labels of j and
m, that is:
g mj ∂k gim = −gim ∂k g mj
nj
g ∂k gin = −gin ∂k g
nj
nj
gmj g ∂k gin = −gmj gin ∂k g
nj
n
δm
∂k gin = −gmj gin ∂k g nj
nj
∂k gim = −gmj gin ∂k g
∂k gij = −gmj gni ∂k g nm
(Eq. 244)
(246)
(m → n)
(247)
(×gmj )
(248)
(Eq. 220)
(249)
(Eq. 172)
(m ↔ j)
(250)
(251)
Similarly, Eq. 243 can be obtained form Eq. 244 by multiplying both sides of Eq. 244
with g in followed by contraction and exchanging the labels of i and n, that is:
in
gim ∂k g mj = −g mj ∂k gim
g gim ∂k g
mj
in mj
= −g g
∂k gim
n
δm
∂k g mj = −g in g mj ∂k gim
∂k g
nj
ij
in mj
= −g g
∂k g = −g
mj in
∂k gim
g ∂k gnm
(Eq. 244)
(252)
in
(×g )
(253)
(Eq. 220)
(254)
(Eq. 172)
(255)
(n ↔ i)
(256)
4.6 Definitions Involving Special Tensors
4.6
97
Definitions Involving Special Tensors
In the following subsections, we investigate a number of mathematical objects and operations whose definitions and applications are dependent on the above described special
tensors, particularly the permutation and metric tensors. These are just a few examples
of tensor definitions involving the special tensors and hence they are not comprehensive
in any way.
4.6.1
Dot Product
The dot product of two basis vectors in general coordinate systems was given earlier in §
4.5 (see Eqs. 213-215). This will be used in the present subsection to develop expressions
for the dot product of vectors and tensors in general.
The dot product of two vectors, A and B, in general coordinate systems using their
covariant and contravariant forms, as well as opposite forms, is given by:
A · B = Ai Ei · Bj Ej = Ai Bj Ei · Ej = g ij Ai Bj = Aj Bj = Ai B i
A · B = Ai Ei · B j Ej = Ai B j Ei · Ej = gij Ai B j = Aj B j = Ai Bi
A · B = Ai Ei · B j Ej = Ai B j Ei · Ej = δ ij Ai B j = Aj B j
A · B = Ai Ei · Bj Ej = Ai Bj Ei · Ej = δij Ai Bj = Ai Bi
(257)
(258)
(259)
(260)
In brief, the dot product of two vectors is the dot product of their two basis vectors
multiplied algebraically by the algebraic product of their components. Because the dot
product of the basis vectors is a metric tensor, the metric tensor will act on the components
by raising or lowering the index of one component or by replacing the index of a component,
as seen in the above equations.
The dot product operations outlined in the previous paragraph can be easily extended
to tensors of higher ranks where the covariant and contravariant forms of the components
and basis vectors are treated in a similar manner to the above examples to obtain the dot
product. For instance, the dot product of a rank-2 tensor of contravariant components Aij
and a vector of covariant components Bk is given by:
(261)
A · B = Aij Ei Ej · Bk Ek = Aij Bk Ei Ej · Ek = Aij Bk Ei δjk = Aij Bj Ei
that is, the ith component of this product, which is a contravariant vector, is:
[A · B]i = Aij Bj
(262)
From the previous statements, we conclude that the dot product in general coordinate
systems occurs between two vectors of opposite variance type. Therefore, to obtain the dot
product of two vectors of the same variance type, one of the vectors should be converted
to the opposite type by the raising or lowering operator, followed by the inner product
operation. This can be generalized to the dot product of higher-rank tensors where the
two contracted indices of the dot product should be of opposite variance type and hence
4.6.2 Magnitude of Vector
98
the index shifting operator in the form of the metric tensor should be used, if necessary,
to achieve this. We note that the generalized dot product of two tensors is invariant under
permissible coordinate transformations. We also note that the variance type of the tensors
involved in an inner product operation is irrelevant for orthonormal Cartesian systems, as
explained before.
4.6.2
Magnitude of Vector
The magnitude of a contravariant vector A is given by:
q
q
p
√
p
p
|A| = A · A = (Ai Ei ) · (Aj Ej ) = (Ei · Ej ) Ai Aj = gij Ai Aj = Aj Aj = Ai Ai
(263)
where Eqs. 213 and 229 are used. A similar expression can be obtained for the covariant
form of the vector, that is:
q
q
p
√
p
p
i
j
|A| = A · A = (Ai E ) · (Aj E ) = (Ei · Ej ) Ai Aj = g ij Ai Aj = Aj Aj = Ai Ai
(264)
where Eqs. 214 and 228 are used. The magnitude of a vector can also be obtained more
directly from the dot product of the covariant and contravariant forms of the vector, that
is:
q
q
q
p
√
p
|A| = A · A = (Ai Ei ) · (Aj Ej ) = (Ei · Ej ) Ai Aj = δji Ai Aj = Ai Ai = Aj Aj
(265)
where Eqs. 215 and 172 are used.
4.6.3
Angle between Vectors
The angle θ between two contravariant vectors A and B is given by:
A·B
Ai Ei · B j Ej
gij Ai B j
=√ k
=p
cos θ =
√
√
|A| |B|
A E k · Al E l B m E m · B n E n
gkl Ak Al gmn B m B n
Similarly, the angle θ between two covariant vectors A and B is given by:
A·B
Ai Ei · Bj Ej
g ij Ai Bj
p
=√
=
√
√
|A| |B|
Ak Ek · Al El Bm Em · Bn En
g kl Ak Al g mn Bm Bn
For two vectors of opposite variance type we have:
cos θ =
A·B
Ai B i
Ai B i
=p
=p
√
√
|A| |B|
gkl Ak Al g mn Bm Bn
g kl Ak Al gmn B m B n
All these three cases can be represented by the following formula:
cos θ =
(266)
(267)
(268)
Aj B j
Ai Bi
√
=
(269)
√
√
Al Al Bn B n
Al Al Bn B n
The angle θ between two sufficiently smooth space curves, C1 and C2 , intersecting at a
given point P in the space is defined as the angle between their tangent vectors, A and
B, at that point (see Figure 17).
cos θ = √
4.6.4 Cross Product
99
B
θ
C1
A
P
C2
Figure 17: The angle between two space curves, C1 and C2 , as the angle θ between their
tangents, A and B, at the point of intersection P .
4.6.4
Cross Product
The cross product of two covariant basis vectors in general coordinate systems of a 3D
space is given by:
Ei × Ej =
∂xm
∂xl ∂xm
∂xl ∂xm
∂xl
e
×
e
=
e
×
e
=
ǫlmn en
l
m
l
m
∂ui
∂uj
∂ui ∂uj
∂ui ∂uj
(270)
where the indexed x and u are the coordinates of Cartesian and general coordinate systems respectively, the indexed e are the Cartesian basis vectors and ǫlmn is the rank-3
permutation relative tensor as defined by Eq. 155. In the last step of the last equation,
Eq. 184 is used to express the cross product of two orthonormal vectors in tensor notation.
We note that for orthonormal Cartesian systems, there is no difference between covariant
and contravariant tensors and hence ei = ei . We also note that for orthonormal Cartesian
systems g = 1 where g is the determinant of the covariant metric tensor (see Eqs. 238 and
236).
n
Now since en = en = ∂x
Ek , the last equation becomes:
∂uk
Ei × Ej =
∂xl ∂xm ∂xn
ǫlmn Ek = ǫijk Ek
∂ui ∂uj ∂uk
(271)
where the underlined absolute covariant permutation tensor is defined as:
ǫijk =
∂xl ∂xm ∂xn
ǫlmn
∂ui ∂uj ∂uk
(272)
So the final result is:
Ei × Ej = ǫijk Ek
(273)
4.6.5 Scalar Triple Product
100
By a similar reasoning, we obtain the following expression for the cross product of two
contravariant basis vectors in general coordinate systems:
Ei × Ej = ǫijk Ek
(274)
where the absolute contravariant permutation tensor is defined as:
ǫijk =
∂ui ∂uj ∂uk lmn
ǫ
∂xl ∂xm ∂xn
(275)
Considering Eq. 166, the above equations can also be expressed as:
Ei × Ej = ǫijk Ek =
√
gǫijk Ek
(276)
ijk
ǫ
Ei × Ej = ǫijk Ek = √ Ek
g
(277)
The cross product of non-basis vectors follows similar rules to those outlined above for
the basis vectors; the only difference is that the algebraic product of the components is
used as a scale factor for the cross product of their basis vectors. For example, the cross
product of two contravariant vectors, Ai and B j , is given by:
(278)
A × B = Ai Ei × B j Ej = Ai B j (Ei × Ej ) = ǫijk Ai B j Ek
that is, the k th component of this product, which is a vector with covariant components,
is:
(279)
[A × B]k = ǫijk Ai B j
Similarly, the cross product of two covariant vectors, Ai and Bj , is given by:
A × B = Ai Ei × Bj Ej = Ai Bj Ei × Ej = ǫijk Ai Bj Ek
(280)
with the k th contravariant component being given by:
[A × B]k = ǫijk Ai Bj
4.6.5
(281)
Scalar Triple Product
The scalar triple product of three contravariant vectors in a 3D space is given by:
(282)
A · (B × C) = Ai Ei · B j Ej × C k Ek
= Ai B j C k Ei · (Ej × Ek )
= Ai B j C k Ei · ǫjkl El
= ǫjkl Ai B j C k Ei · El
(Eq. 273)
= ǫjkl Ai B j C k δil
(Eq. 215)
= ǫjki Ai B j C k
(Eq. 172)
4.6.6 Vector Triple Product
101
= ǫijk Ai B j C k
(Eq. 180)
Similarly, the scalar triple product of three covariant vectors in a 3D space is given by:
(283)
A · (B × C) = Ai Ei · Bj Ej × Ck Ek
= Ai Bj Ck Ei · Ej × Ek
(Eq. 274)
= Ai Bj Ck Ei · ǫjkl El
i
jkl
= ǫ Ai B j C k E · E l
= ǫjkl Ai Bj Ck δli
(Eq. 215)
jki
(Eq. 172)
ijk
(Eq. 180)
= ǫ Ai B j C k
= ǫ Ai B j C k
Other forms can also be obtained by a similar method.
4.6.6
Vector Triple Product
The vector triple product of three vectors, one contravariant and two covariant, in a 3D
space is given by:
(284)
A × (B × C) = Ai Ei × Bj Ej × Ck Ek
= Ai B j C k E i × E j × E k
(Eq. 274)
= Ai Bj Ck Ei × ǫjkl El
= ǫjkl Ai Bj Ck (Ei × El )
= ǫjkl Ai Bj Ck ǫilm Em
(Eq. 273)
= ǫilm ǫjkl Ai Bj Ck Em
= ǫilm ǫjkl Ai Bj Ck Em
(Eq. 166)
Following relabeling of indices and writing in a covariant component form, we obtain the
following more organized expression:
[A × (B × C)]i = ǫijk ǫklm Aj Bl Cm
(285)
We can also obtain a different form from one covariant vector and two contravariant
vectors, that is:
(286)
A × (B × C) = Ai Ei × B j Ej × C k Ek
= Ai B j C k Ei × (Ej × Ek )
= Ai B j C k Ei × ǫjkl El
= ǫjkl Ai B j C k Ei × El
= ǫjkl Ai B j C k ǫilm Em
(Eq. 273)
(Eq. 274)
4.6.7 Determinant of Matrix
102
= ǫilm ǫjkl Ai B j C k Em
= ǫilm ǫjkl Ai B j C k Em
(Eq. 166)
Following relabeling of indices and writing in a contravariant component form, we obtain
the following more organized expression:
[A × (B × C)]i = ǫijk ǫklm Aj B l C m
(287)
The expressions of the other principal form of the vector triple product, i.e. (A × B) × C,
can be obtained by a similar method.
4.6.7
Determinant of Matrix
For a 3 × 3 matrix representing a rank-2 tensor A of mixed form in a 3D space, the
determinant is given by:
det (A) =
A11 A12 A13
A21 A22 A23
A31 A32 A33
=
1 ijk
1 ijk l m n
n
ǫ ǫlmn Ali Am
δ AA A
j Ak =
3!
3! lmn i j k
(288)
where Eq. 201 is used in the last step. This may be expressed as a row expansion by
substituting 1, 2, 3 for l, m, n that is:
det (A) = ǫijk A1i A2j A3k
(289)
It can also be expressed as a column expansion by substituting 1, 2, 3 for i, j, k that is:
n
det (A) = ǫlmn Al1 Am
2 A3
(290)
More generally, for an n × n matrix representing a rank-2 tensor A of mixed form in an
nD space, the determinant is given by:
1 i1 ···in
ǫj1 ···jn Aji11 . . . Ajinn
(291)
ǫ
n!
Similar definitions can be given using the covariant and contravariant forms of the tensor
with the employment of the opposite variance type of the permutation tensors.
det (A) =
4.6.8
Length
The differential of displacement vector in general coordinate systems is given by:
X
Ei i X
∂r i
i
du
=
E
du
=
|E
|
du =
|Ei | Êi dui
dr =
i
i
i
∂u
|Ei |
i
i
(292)
where r is the position vector as defined previously and the hat indicates a normalized
vector. The length of line element, ds, which may also be called the differential of arc
length, in general coordinate systems is given by:
(ds)2 = dr · dr = Ei dui · Ej duj = (Ei · Ej ) dui duj = gij dui duj
(293)
4.6.9 Area
103
where gij is the covariant metric tensor.
For orthogonal coordinate systems, the metric tensor is given by:
(
0
(i 6= j)
gij =
2
(hi )
(i = j)
(294)
where hi is the scale factor of the ith coordinate. Hence, the last part of Eq. 293 becomes:
X
(ds)2 =
(hi )2 dui dui
(295)
i
with no cross terms (i.e. terms of products involving more than one coordinate like dui duj
where i 6= j) which are generally present in the case of non-orthogonal coordinate systems.
On conducting a transformation from one coordinate system to another coordinate system, where the other system is marked with barred coordinates ū, the length of line element
ds will be expressed in the new system as:
(ds)2 = ḡij dūi dūj
(296)
Since the length of line element is an invariant quantity, the same symbol (ds)2 is used in
both Eqs. 293 and 296.
Based on the above formulations, the length L of a t-parameterized space curve C(t)
defined by ui = ui (t) where i = 1, · · · , n, which represents the distance traversed along the
curve on moving between its start point P1 and end point P2 , is given in general coordinate
systems of an nD space by:
ˆ t2 r
ˆ P2 p
ˆ
dui duj
i
j
dt
(297)
gij du du =
gij
ds =
L=
dt dt
P1
C
t1
where C represents the space curve, t is a scalar real variable, and t1 and t2 are the values
of t corresponding to the start and end points respectively. It is noteworthy that some
authors add a sign indicator to ensure that the argument of the square root in the above
equation is positive. However, as indicated in § 1.1, such a condition is assumed when
needed since in this book we deal with real quantities only.
4.6.9
Area
In general coordinate systems of a 3D space, the area of an infinitesimal element of the
surface u1 = c1 in the neighborhood of a given point P , where c1 is a constant, is obtained
by taking the magnitude of the cross product of the differentials of the displacement vectors
in the directions of the other two coordinates on that surface at P , i.e. the tangent vectors
to the other two coordinate curves at P (see Figure 18). Hence, the area of a differential
element on the surface u1 = c1 is given by:
dσ(u1 = c1 ) = |dr2 × dr3 |
(298)
4.6.9 Area
104
∂r
∂r
× 3 du2 du3
2
∂u
∂u
= |E2 × E3 | du2 du3
=
1
2
= ǫ231 E du du
(Eq. 45)
3
(Eq. 273)
= |ǫ231 | E1 du2 du3
√ √
= g E1 · E1 du2 du3
√ p
= g g 11 du2 du3
p
= gg 11 du2 du3
(Eqs. 166 & 263)
(Eq. 214)
where σ represents area and dr2 and dr3 are the displacement differentials along the second
and third coordinate curves at P while the other symbols are as defined previously.
dr2 × dr3
dr3
dσ
u2
P
dr2
u3
Figure 18: The area dσ of an infinitesimal element of the surface u1 = c1 in the neighborhood of a given point P as the magnitude of the cross product of the differentials of the
displacement vectors in the directions of the other two coordinate curves on that surface
at P , dr2 and dr3 .
On generalizing the above argument, the area of a differential element on the surface
ui = ci (i = 1, 2, 3) in a 3D space where ci is a constant is given by:
p
(i 6= j 6= k, no sum on i)
(299)
dσ(ui = ci ) = gg ii duj duk
In orthogonal coordinate systems in a 3D space we have:
s
p
1
= hj hk
(i 6= j 6= k, no sum on any index) (300)
gg ii = (hi )2 (hj )2 (hk )2
(hi )2
where Eqs. 235 and 236 are used. Hence, Eq. 299 becomes:
dσ(ui = ci ) = hj hk duj duk
(i 6= j 6= k, no sum on any index)
(301)
4.6.10 Volume
105
The last formula represents the area of a surface element of a rectangular shape with sides
hj duj and hk duk (no sum on j or k).
Based on the above discussion, the area A of a finite surface patch can be defined by the
following formula:
¨
A=
dσ
(302)
S
where S represents a surface patch and dσ is the area differential of an infinitesimal
element on the patch. For coordinate surfaces, the expression for dσ can be obtained from
the previous formulations.
4.6.10
Volume
In general coordinate systems of a 3D space, the volume of an infinitesimal element of
a solid body occupying a given region of the space in the neighborhood of a given point
P , where the element is represented by a parallelepiped defined by the three differentials
of the displacement vectors dri (i = 1, 2, 3) along the three coordinate curves at P , is
obtained by taking the magnitude of the scalar triple product of these vectors (refer to §
4.6.5 and see Figure 19). Hence, the volume of a differential element of the body is given
by:
dτ = |dr1 · (dr2 × dr3 )|
∂r
∂r
∂r
·
× 3 du1 du2 du3
=
1
2
∂u
∂u
∂u
= |E1 · (E2 × E3 )| du1 du2 du3
1
1
2
= E1 · ǫ231 E du du du
3
= E1 · E1 |ǫ231 | du1 du2 du3
= δ11 |ǫ231 | du1 du2 du3
√
= g du1 du2 du3
= J du1 du2 du3
(303)
(Eq. 45)
(Eq. 273)
(Eq. 215)
(Eq. 166)
(Eq. 63)
where g is the determinant of the covariant metric tensor gij , and J is the Jacobian of
the transformation as defined previously (see § 2.3). We note that due to the freedom
of choice of the order of the variables, which is related to the choice of the coordinate
system handedness that could affect the sign of the determinant Jacobian, the sign of the
determinant should be adjusted if necessary to have a proper sign for the volume element
as a positive quantity. The last line in the last equation is particularly useful for changing
the variables in multivariate integrals where the Jacobian facilitates the transformation.
In orthogonal coordinate systems in a 3D space, the above formulation becomes:
dτ = h1 h2 h3 du1 du2 du3
(304)
where h1 , h2 and h3 are the scale factors and where Eq. 236 is used. Geometrically, the
last formula represents the volume of a rectangular parallelepiped with edges h1 du1 , h2 du2
and h3 du3 .
4.6.10 Volume
106
dr3
dr2
dr1
u1
P
u2
u3
Figure 19: The volume of an infinitesimal element of a solid body in a 3D space in the
neighborhood of a given point P as the magnitude of the scalar triple product of the
differentials of the displacement vectors in the directions of the three coordinate curves at
P , dr1 , dr2 and dr3 .
The formulae of Eq. 303, which are specific to a 3D space, can be extended to the
differential of a generalized volume element in general coordinate systems of an nD space
as follows:
√
(305)
dτ = gdu1 . . . dun = J du1 . . . dun
We note that generalized volume elements are used, for instance, to represent the change
of variables in multi-variable integrations.
Based on the above discussion, the volume V of a solid body occupying a finite region
of the space can be defined as:
˚
V =
dτ
(306)
Ω
where Ω represents the region of the space occupied by the solid body and dτ is the volume
differential of an infinitesimal element of the body, and where the expression for dτ should
be obtained from the above formulations.
4.7 Exercises and Revision
4.7
107
Exercises and Revision
4.1 What is special about the Kronecker delta, the permutation and the metric tensors
and why they deserve special attention?
4.2 Give detailed definition, in words and in symbols, of the ordinary Kronecker delta
tensor in an nD space.
4.3 List and discuss all the main characteristics (e.g. symmetry) of the ordinary Kronecker
delta tensor.
4.4 Write the matrix that represents the ordinary Kronecker delta tensor in a 4D space.
4.5 Do we violate the rules of tensor indices when we write: δij = δ ij = δ ij = δij ?
4.6 Explain the following statement: “The ordinary Kronecker delta tensor is conserved
under all proper and improper coordinate transformations”. What is the relation
between this and the property of isotropy of this tensor?
4.7 List and discuss all the main characteristics (e.g. anti-symmetry) of the permutation
tensor.
4.8 What are the other names used to label the permutation tensor?
4.9 Why the rank and the dimension of the permutation tensor are the same? Accordingly,
what is the number of components of the rank-2, rank-3 and rank-4 permutation
tensors?
4.10 Why the permutation tensor of any rank has only one independent non-vanishing
component?
4.11 Prove that the rank-n permutation tensor possesses n! non-zero components.
4.12 Why the permutation tensor is totally anti-symmetric?
4.13 Give the inductive mathematical definition of the components of the permutation
tensor of rank-n.
4.14 State the most simple analytical mathematical definition of the components of the
permutation tensor of rank-n.
4.15 Make a sketch of the array representing the rank-3 permutation tensor where the
nodes of the array are marked with the symbols and values of the components of this
tensor.
4.16 Define, mathematically, the rank-n covariant and contravariant absolute permutation
tensors, ǫi1 ...in and ǫi1 ...in .
4.17 Show that ǫijk is a relative tensor of weight −1 and ǫijk is a relative tensor of weight
+1.
√
4.18 Show that ǫi1 ...in = g ǫi1 ...in and ǫi1 ...in = √1g ǫi1 ...in are absolute tensors assuming that
ǫi1 ...in is a relative tensor of weight −1 and ǫi1 ...in is a relative tensor of weight +1.
4.19 Write ǫi1 i2 ···in ǫj1 j2 ···jn in its determinantal form in terms of the ordinary Kronecker
delta.
4.20 Prove the following identity: ǫi1 i2 ···in ǫi1 i2 ···in = n!.
4.21 State a mathematical relation representing the use of the ordinary Kronecker delta
tensor acting as an index replacement operator.
4.22 Prove the following relation inductively by writing it in an expanded form in a 3D
space: δii = n.
4.7 Exercises and Revision
108
4.23 Repeat exercise 4.22 with the relation: ui,j = δji using a matrix form.
4.24 Justify the following relation assuming an orthonormal Cartesian system: ∂i xj = ∂j xi .
4.25 Justify the following relations where the indexed e are orthonormal vectors:
ei · ej = δij
ei ej : ek el = δik δjl
4.26 Show that δij δjk δki = n.
4.27 Write the determinantal array form of ǫij ǫkl outlining the pattern of the tensor indices
in their relation to the indices of the rows and columns of the determinant array.
4.28 Prove the following identity using a truth table: ǫij ǫkl = δki δlj − δli δkj .
4.29 Prove the following identities justifying each step in your proofs:
ǫil ǫkl = δki
j
i j
ǫijk ǫlmk = δli δm
− δm
δl
4.30 Prove the following identities using other more general identities:
ǫijk ǫljk = 2δli
ǫijk ǫijk = 6
4.31 Outline the similarities and differences between the ordinary Kronecker delta tensor
and the generalized Kronecker delta tensor.
4.32 Give the inductive mathematical definition of the generalized Kronecker delta tensor
n
δji11 ...i
...jn .
n
4.33 Write the determinantal array form of the generalized Kronecker delta tensor δji11 ...i
...jn
in terms of the ordinary Kronecker delta tensor.
4.34 Define ǫi1 ...in and ǫi1 ...in in terms of the generalized Kronecker delta tensor.
ijk
using an analytic or an inductive or a truth table
4.35 Prove the relation: ǫijk ǫlmn = δlmn
method.
4.36 Demonstrate that the generalized Kronecker delta is an absolute tensor.
mnq
mn
= δkl
.
4.37 Prove the following relation justifying each step in your proof: δklq
4.38 Prove the common form of the epsilon-delta identity.
4.39 Prove the following generalization of the epsilon-delta identity: g ij ǫikl ǫjmn = gkm gln −
gkn glm .
4.40 List and discuss all the main characteristics (e.g. symmetry) of the metric tensor.
4.41 How many types the metric tensor has?
4.42 Investigate the relation of the metric tensor of a given space to the coordinate systems
of the space as well as its relation to the space itself by comparing the characteristics
of the metric in different coordinate systems of the space such as being diagonal or
not or having constant or variable components and so on. Hence, assess the status of
the metric as a property of the space but with a form determined by the adopted coordinate system to describe the space and hence it is also a property of the coordinate
system in this sense.
4.43 What is the relation between the covariant metric tensor and the length of an infinitesimal element of arc ds in a general coordinate system?
4.44 How the relation in question 4.43 will become (a) in an orthogonal coordinate system
and (b) in an orthonormal Cartesian coordinate system?
4.7 Exercises and Revision
109
4.45 What is the characteristic feature of the metric tensor in orthogonal coordinate systems?
4.46 Write the mathematical expressions for the components of the covariant, contravariant
and mixed forms of the metric tensor in terms of the covariant and contravariant basis
vectors, Ei and Ei .
4.47 Write, in full tensor notation, the mathematical expressions for the components of the
covariant and contravariant forms of the metric tensor, gij and g ij .
4.48 What is the relation between the mixed form of the metric tensor and the ordinary
Kronecker delta tensor?
4.49 Explain why the metric tensor is not necessarily diagonal in general coordinate systems
but it is necessarily symmetric.
4.50 Explain why the diagonal elements of the metric tensor in general coordinate systems
are not necessarily of unit magnitude or positive but they are necessarily non-zero.
4.51 Explain why the mixed type metric tensor in any coordinate system is diagonal or in
fact it is the unity tensor.
4.52 Show that the covariant and contravariant forms of the metric tensor, gij and g ij , are
inverses of each other.
4.53 Why the determinant of the metric tensor should not vanish at any point in the space?
4.54 If the determinant of the covariant metric tensor gij is g, what is the determinant of
the contravariant metric tensor g ij ?
4.55 Show that the metric tensor can be regarded as a transformation of the ordinary
Kronecker delta tensor in its different variance types from an orthonormal Cartesian
coordinate system to a general coordinate system.
4.56 Justify the use of the metric tensor as an index shifting operator using a mathematical
argument.
4.57 Carry out the following index shifting operations recording the order of the indices:
g ij Cklj
gmn B nst
gnl Dkmn
4.58 What is the difference between the three operations in question 4.57?
4.59 Why the order of the raised and lowered indices is important and hence it should be
recorded? Mention one form of notation used to record the order of the indices.
4.60 What is the condition that should be satisfied by the metric tensor of a flat space?
Give common examples of flat and curved spaces.
4.61 Considering a coordinate transformation, what is the relation between the determinants of the covariant metric tensor in the original and transformed coordinate
systems, g and ḡ?
4.62 B is a “conjugate” or “associate” tensor of tensor A. What this means?
4.63 Complete and justify the following statement: “The components of the metric tensor
are constants iff ...etc.”.
4.64 What are the covariant and absolute derivatives of the metric tensor?
4.65 Assuming an orthogonal coordinate system of an nD space, complete the following
equations where the indexed g represents the metric tensor or its components, i 6= j
4.7 Exercises and Revision
110
in the second equation and there is no sum in the third equation:
gij =?
g ii =?
det g ij =?
4.66 Write, in matrix form, the covariant and contravariant metric tensor for orthonormal
Cartesian, cylindrical and spherical coordinate systems. What distinguishes all these
matrices? Explain and justify.
4.67 Referring to question 4.66, what is the relation between the diagonal elements of these
matrices and the scale factors hi of the coordinates of these systems?
4.68 Considering the Minkowski metric, is the space of the mechanics of Lorentz transformations flat or curved? Is it homogeneous or not? What effect this can have on the
length of element of arc ds?
4.69 Derive the following identities:
gim ∂k g mj = −g mj ∂k gim
∂k gij = −gmj gni ∂k g nm
4.70 What is the dot product of A and B where A is a rank-2 covariant tensor and B is a
contravariant vector? Write this operation in steps providing full justification of each
step.
4.71 Derive an expression for the magnitude of a vector A when A is covariant and when
A is contravariant.
4.72 Derive an expression for the cosine of the angle θ between two covariant vectors, A
and B, and between two contravariant vectors C and D.
4.73 What is the meaning of the angle between two intersecting smooth curves?
4.74 What is the cross product of A and B where these are covariant vectors?
4.75 Complete the following equations assuming a general coordinate system of a 3D space:
Ei × Ej =?
Ei × Ej =?
4.76 Define the operations of scalar triple product and vector triple product of vectors
using tensor language and assuming a general coordinate system of a 3D space.
4.77 What is the relation between the relative and absolute permutation tensors in their
covariant and contravariant forms?
4.78 Define the determinant of a matrix B in tensor notation assuming a general coordinate
system of a 3D space.
4.79 Derive the relation for the length of line element in general coordinate systems: (ds)2 =
gij dui duj . How will this relation become when the coordinate system is orthogonal?
Justify your answer.
4.80 Write the integral representing the length L of a t-parameterized space curve in terms
of the metric tensor.
4.81 Using Eq. 295 plus the scale factors of Table 1, develop expressions for ds in orthonormal Cartesian, cylindrical and spherical coordinate systems.
4.82 Derive the following formula for the area of a differential element on the coordinate
surface ui = constant in a 3D space assuming a general coordinate system:
p
dσ(ui = constant) = gg ii duj duk
(i 6= j 6= k, no sum on i)
4.7 Exercises and Revision
111
How will this relation become when the coordinate system is orthogonal?
4.83 Using Eq. 301 plus the scale factors of Table 1, develop expressions for dσ on the
coordinate surfaces in orthonormal Cartesian, cylindrical and spherical coordinate
systems.
4.84 Derive the following formula for the volume of a differential element of a solid body
in a 3D space assuming a general coordinate system:
dτ =
√
g du1 du2 du3
How will this relation become when the coordinate system is orthogonal?
4.85 Make a plot representing the volume of an infinitesimal element of a solid body in a
3D space as the magnitude of a scalar triple product of three vectors.
4.86 Use the expression of the volume element in general coordinate systems of nD spaces
to find the formula for the volume element in orthogonal coordinate systems.
4.87 Using Eq. 304 plus the scale factors of Table 1, develop expressions for dτ in orthonormal Cartesian, cylindrical and spherical coordinate systems.
Chapter 5
Tensor Differentiation
Ordinary differentiation rules of partial and total derivatives do not satisfy the principle
of invariance when they are applied to tensors in general coordinate systems due to the
fact that the basis vectors in these systems are coordinate dependent. This means that
the ordinary differentiation of non-scalar tensor components in general coordinates does
not necessarily results in a tensor. Therefore, special types of differentiation should be
defined so that when they apply to tensors they produce tensors. The essence of these
operations is to extend the differentiation process to the basis vectors to which the tensor
is referred and not only on the tensor components.
The focus of this chapter is the operations of covariant and absolute differentiation which
are closely linked. These operations represent generalization of ordinary differentiation in
general coordinate systems with an objective of making the derivative of tensors comply
with the principle of tensor invariance. Briefly, the differential change of a tensor in
general coordinate systems is the result of a change in the basis vectors and a change
in the tensor components. Hence, covariant and absolute differentiation, in place of the
normal differentiation (i.e. partial and total differentiation respectively), are defined and
employed to account for both of these changes by differentiating the basis vectors as well
as the components of the tensors.
Since the Christoffel symbols are crucial in the formulation and application of covariant
and absolute differentiation, the first section of the present chapter is dedicated to these
symbols and their properties. The subsequent two sections will then focus on the covariant
differentiation and the absolute differentiation.
5.1
Christoffel Symbols
We start by investigating the main properties of the Christoffel symbols which play crucial
roles in tensor calculus in general and are needed for the subsequent development of the
upcoming sections since they enter in the definition of covariant and absolute differentiation. The Christoffel symbols are classified as those of the first kind and those of the
second kind. These two kinds are linked through the index raising and lowering operators
although this does not mean they are general tensors (see next). Both kinds of Christoffel
symbols are variable functions of coordinates since they depend in their definition on the
metric tensor which is coordinate dependent in general.
The Christoffel symbols of the first and second kind are not general tensors although
they are affine tensors of rank-3. We note that affine tensors (see § 3.1.6) are tensors that
correspond to admissible linear coordinate transformations from an original rectangular
system of coordinates. As a consequence of not being tensors, if all the Christoffel symbols
of either kind vanish in a particular coordinate system they will not necessarily vanish in
112
5.1 Christoffel Symbols
113
other coordinate systems (see § 3.1.4). For instance, all the Christoffel symbols of both
kinds vanish in Cartesian coordinate systems but not in cylindrical or spherical coordinate
systems, as will be established later in this section.
The Christoffel symbols of the first kind are defined as:
[ij, l] =
1
(∂j gil + ∂i gjl − ∂l gij )
2
(307)
where the indexed g is the covariant form of the metric tensor. The Christoffel symbols of
the second kind are obtained by raising the third index of the Christoffel symbols of the
first kind, and hence they are given by:
Γkij = g kl [ij, l] =
g kl
(∂j gil + ∂i gjl − ∂l gij )
2
(308)
where the indexed g is the metric tensor in its contravariant and covariant forms with
implied summation over l. Similarly, the Christoffel symbols of the first kind can be
obtained from the Christoffel symbols of the second kind by reversing the above process
through lowering the upper index, that is:
l
gkm Γkij = gkm g kl [ij, l] = δm
[ij, l] = [ij, m]
(309)
where Eqs. 308, 220 and 172 are used.
The Christoffel symbols of the first and second kind are symmetric in their paired indices,
that is:
[ij, k] = [ji, k]
Γkij = Γkji
(310)
(311)
These properties can be verified by shifting the indices in the definitions of the Christoffel
symbols, as given by Eqs. 307 and 308, noting that the metric tensor is symmetric in its
two indices (see § 4.5).
For an nD space with n covariant basis vectors (E1 , E2 , . . . , En ) spanning the space, the
partial derivative ∂j Ei for any given i and j is a vector within the space and hence it is in
general a linear combination of all the basis vectors. The Christoffel symbols of the second
kind are the components of this linear combination, that is:
∂j Ei = Γkij Ek
(312)
Similarly, for the contravariant basis vectors (E1 , E2 , . . . , En ) we have:
∂j Ei = −Γikj Ek
(313)
By inner product multiplication of Eq. 312 with Ek and Eq. 313 with Ek we obtain:
Ek · ∂j Ei = +Γkij Ek · Ek = +Γkij δkk = +Γkij
(314)
5.1 Christoffel Symbols
114
Ek · ∂j Ei = −Γikj Ek · Ek = −Γikj δkk = −Γikj
(315)
where Eqs. 49 and 172 are used. These equations reveal that the Christoffel symbols of
the second kind are the projections of the partial derivative of the basis vectors in the
direction of the basis vectors of the opposite variance type. From Eq. 314 and by using
the index lowering operator, we obtain a similar relation that links the Christoffel symbols
of the first kind to the basis vectors and their partial derivatives, that is:
Ek · ∂j Ei = gmk Em · ∂j Ei = gmk Γm
ij = [ij, k]
(316)
where Eq. 309 is used in the last step. This equation reveals that the Christoffel symbols
of the first kind are the projections of the partial derivative of the covariant basis vectors
in the direction of the basis vectors of the same variance type.
The partial derivative of the components of the covariant metric tensor and the Christoffel
symbols of the first kind satisfy the following identity:
(317)
∂k gij = [ik, j] + [jk, i]
This relation can be obtained from the partial derivative of the dot product of the basis
vectors with the use of Eq. 316, that is:
∂k gij = ∂k (Ei · Ej )
= (∂k Ei ) · Ej + Ei · (∂k Ej )
= [ik, j] + [jk, i]
(Eq. 213)
(product rule)
(Eq. 316)
(318)
We note that Eq. 317 is closely linked to the upcoming Ricci theorem (see § 5.2), that is:
gij;k = ∂k gij −gaj Γaik −gia Γajk = ∂k gij −[ik, j]−[jk, i] = 0 ⇐⇒ ∂k gij = [ik, j]+[jk, i] (319)
The relation given by Eq. 317 can also be written in terms of the Christoffel symbols of
the second kind using the index shifting operator, that is:
∂k gij = gaj Γaik + gai Γajk
(320)
where Eq. 309 is used.
Following the method of derivation given in Eq. 318, we obtain the following relation
for the partial derivative of the components of the contravariant metric tensor:
(Eq. 214)
(321)
∂k g ij = ∂k Ei · Ej
(product rule)
= ∂ k Ei · Ej + Ei · ∂ k Ej
j
j
a
i
i
a
(Eq. 313)
= −Γak E · E + E · −Γak E
= −Γiak Ea · Ej − Γjak Ei · Ea
= −g aj Γiak − g ia Γjak
(Eq. 214)
5.1 Christoffel Symbols
that is:
115
∂k g ij = −g aj Γiak − g ia Γjak
(322)
g ij;k = ∂k g ij + g aj Γiak + g ia Γjak = 0 ⇐⇒ ∂k g ij = −g aj Γiak − g ia Γjak
(323)
Like Eq. 317 whose close link to the Ricci theorem is shown in Eq. 319, Eq. 322 can
also be seen from this perspective, that is:
The Christoffel symbols of the second kind with two identical indices of opposite variance
type satisfy the following relations:
Γjji = Γjij =
1 √
1
1
√
∂i g = ∂i (ln g) = ∂i (ln g) = √ ∂i g
2g
2
g
(324)
where the main relation can be derived as follows:
Γjij =
=
=
=
=
g jl
(∂j gil + ∂i gjl − ∂l gij )
2
g jl
(∂l gij + ∂i gjl − ∂l gij )
2
1 jl
g ∂i gjl
2
1 jl
gg ∂i gjl
2g
1
∂i g
2g
(Eq. 308 with k = j)
(325)
(relabeling dummy j, l in 1st term & g jl = g lj )
(derivative of determinant)
All the other forms of Eq. 324 can be obtained from the derived form by simple algebraic
manipulations with the use of the rules of differentiation and natural logarithms.
In orthogonal coordinate systems, the Christoffel symbols of the first kind can be classified into three main groups, considering the identicality and difference of their indices,
which are given by:
1
[ij, i] = [ji, i] = ∂j gii
2
1
[ii, j] = − ∂j gii
2
[ij, k] = 0
(no sum on i)
(326)
(i 6= j, no sum on i)
(327)
(i 6= j 6= k)
(328)
These three equations can be obtained directly from Eq. 307 with proper labeling of the
indices (i.e. according to the labeling of the left hand side of Eqs. 326-328) noting that for
orthogonal coordinate systems gij = 0 when i 6= j. For example, Eq. 327 can be obtained
as follows:
1
(∂i gij + ∂i gij − ∂j gii )
2
1
= (0 + 0 − ∂j gii )
2
[ii, j] =
(Eq. 307 with j → i and l → j)
(gij = 0)
(329)
5.1 Christoffel Symbols
116
1
= − ∂j gii
2
In fact, Eq. 326 can be obtained from Eq. 317 by exchanging j and k then replacing k
with i followed by a simple algebraic manipulation. The middle equality of Eq. 326 is
based on the fact that the Christoffel symbols are symmetric in their paired indices (Eq.
310). We note that Eq. 326 includes the case of j = i, i.e. when all the three indices are
identical, and hence [ii, i] = 21 ∂i gii . For this reason, we did not add the condition i 6= j to
Eq. 326.
Returning to the above indicated consideration of the identicality and difference of the
indices of the Christoffel symbols of the first kind, we have 4 main cases: (1) all the indices
are identical, (2) only two non-paired indices are identical, (3) only the two paired indices
are identical, and (4) all the indices are different. Eq. 326 represents case 1 and case 2,
Eq. 327 represents case 3, and Eq. 328 represents case 4. This classification similarly
applies to the Christoffel symbols of the second kind as represented by the upcoming Eqs.
331-333.
In orthogonal coordinate systems where gij = g ij = 0 (i 6= j), the Christoffel symbols of
the second kind are given by:
Γijk = g ii [jk, i]
[jk, i]
=
gii
(Eq. 308)
(330)
(Eq. 235)
with no sum on i. Hence, from Eqs. 326-328 and Eq. 330 the Christoffel symbols of the
second kind in orthogonal coordinate systems are given by:
g ii
1
1
=
=
∂j gii =
∂j gii = ∂j ln gii
2
2gii
2
jj
1
g
∂j gii
Γjii = − ∂j gii = −
2
2gjj
Γijk = 0
Γiij
Γiji
(no sum on i)
(331)
(i 6= j, no sum on i or j)
(332)
(i 6= j 6= k)
(333)
where in the last step of Eq. 331 the well known rule of differentiating the natural logarithm
is used. Like the Christoffel symbols of the first kind (refer to Eq. 326), Eq. 331 also
includes the case of j = i, i.e. when all the three indices are identical, that is Γiii = 2g1ii ∂i gii .
In orthogonal coordinate systems of a 3D space, the Christoffel symbols of both kinds
vanish when the indices are all different, as seen above (Eqs. 328 and 333). Hence, out
of a total of 27 symbols, representing all the possible permutations of the three indices
including the repetitive ones, only 21 non-identically vanishing symbols are left since the six
non-repetitive permutations are dropped. If we now consider that the Christoffel symbols
are symmetric in their paired indices (Eqs. 310 and 311), then we are left with only 15
independent non-identically vanishing symbols since six other permutations representing
these symmetric exchanges are also dropped because they are not independent.
5.1 Christoffel Symbols
117
Accordingly, in orthogonal coordinate systems in a 3D space the 15 independent nonidentically vanishing Christoffel symbols of the first kind are:
[11, 1] = +h1 h1,1
[12, 1] = +h1 h1,2
[13, 3] = +h3 h3,1
[22, 3] = −h2 h2,3
[33, 1] = −h3 h3,1
[11, 2] = −h1 h1,2
[12, 2] = +h2 h2,1
[22, 1] = −h2 h2,1
[23, 2] = +h2 h2,3
[33, 2] = −h3 h3,2
[11, 3] = −h1 h1,3
[13, 1] = +h1 h1,3
[22, 2] = +h2 h2,2
[23, 3] = +h3 h3,2
[33, 3] = +h3 h3,3
(334)
where the indices 1, 2, 3 stand for the three coordinates, h1 , h2 , h3 are the scale factors
corresponding to these coordinates as defined previously (see § 2.5 and 2.6), and the
comma indicates, as always, partial derivative with respect to the coordinate represented
by the following index. For example, in cylindrical coordinates given by (ρ, φ, z), h2,1 means
the partial derivative of h2 with respect to the first coordinate and hence h2,1 = ∂ρ ρ = 1
since h2 = ρ (refer to Table 1) and the first coordinate is ρ. Similarly, in spherical
coordinates given by (r, θ, φ), h3,2 means the partial derivative of h3 with respect to the
second coordinate and hence h3,2 = ∂θ (r sin θ) = r cos θ since h3 = r sin θ (refer to Table
1) and the second coordinate is θ. As indicated above, because the Christoffel symbols of
the first kind are symmetric in their first two indices, the expression of the [21, 1] symbol
for instance can be obtained from the expression of the [12, 1] symbol.
The expressions given in Eq. 334 for the Christoffel symbols of the first kind in orthogonal
coordinate systems are no more than simple applications of Eqs. 326 and 327 plus Eq.
235. For example, the entry [12, 1] can be obtained as follows:
1
[12, 1] = ∂2 g11
2
1
= ∂2 (h1 )2
2
= h1 ∂ 2 h1
= h1 h1,2
(Eq. 326)
(335)
(Eq. 235)
(rules of differentiation)
(notation)
Similarly, in orthogonal coordinate systems in a 3D space the 15 independent nonidentically vanishing Christoffel symbols of the second kind are:
h1,1
h1
h1,2
Γ112 = +
h1
h
3,1
Γ313 = +
h3
h2 h2,3
Γ322 = −
(h3 )2
h3 h3,1
Γ133 = −
(h1 )2
Γ111 = +
h1 h1,3
(h3 )2
h1,3
=+
h1
h2,2
=+
h2
h3,2
=+
h3
h3,3
=+
h3
Γ211 = − h(h1 h1,2
)2
Γ311 = −
Γ212 = + hh2,1
2
Γ113
Γ122 = − h(h2 h2,1
)2
1
Γ222
Γ223 = + hh2,3
2
Γ323
Γ233 = − h(h3 h3,2
)2
Γ333
2
2
(336)
5.1 Christoffel Symbols
118
where the symbols are as explained above. Again, since the Christoffel symbols of the
second kind are symmetric in their lower indices, the non-vanishing entries which are not
listed above can be obtained from the given entries by permuting the lower indices. The
relations of Eq. 336 can be obtained from the relations of Eq. 334 by dividing by (hi )2
where i is the third index of the Christoffel symbol of the first kind. This can be justified
by Eqs. 330 and 235. The relations of Eq. 336 can also be obtained directly from Eqs.
331 and 332 plus Eq. 235, as done for the Christoffel symbols of the first kind where Eqs.
326 and 327 were used.
In any coordinate system, all the Christoffel symbols of the first and second kind vanish
identically iff all the components of the metric tensor in the given coordinate system are
constants. This can be seen from the definitions of the Christoffel symbols, as given by
Eqs. 307 and 308, since the partial derivatives will vanish in this case. In affine coordinate
systems all the components of the metric tensor are constants and hence all the Christoffel
symbols of both kinds vanish identically. The prominent example is the orthonormal
Cartesian coordinate systems where all the Christoffel symbols of the first and second
kind are identically zero. This can also be seen from Eqs. 334 and 336 since the scale
factors are constants (refer to Table 1).
In cylindrical coordinate systems, identified by the coordinates (ρ, φ, z), all the Christoffel
symbols of the first kind are zero except:
[22, 1] = −ρ
[12, 2] = [21, 2] = ρ
(337)
(338)
where the indices 1, 2 stand for the coordinates ρ, φ respectively. Also, in cylindrical
coordinate systems the non-zero Christoffel symbols of the second kind are given by:
Γ122 = −ρ
(339)
1
(340)
Γ212 = Γ221 =
ρ
where the symbols are as explained above. Again, these results can be obtained from Eqs.
334 and 336 using the scale factors of Table 1.
In spherical coordinate systems, identified by the coordinates (r, θ, φ), the non-zero
Christoffel symbols of the first kind are given by:
[22, 1]
[33, 1]
[12, 2]
[33, 2]
[13, 3]
= −r
= −r sin2 θ
= [21, 2] = r
= −r2 sin θ cos θ
= [31, 3] = r sin2 θ
[23, 3] =
[32, 3] = r2 sin θ cos θ
(341)
(342)
(343)
(344)
(345)
(346)
where the indices 1, 2, 3 stand for the coordinates r, θ, φ respectively. Also, in spherical
coordinate systems the non-zero Christoffel symbols of the second kind are given by:
Γ122 = −r
(347)
5.1 Christoffel Symbols
119
Γ133 = −r sin2 θ
1
r
= − sin θ cos θ
1
= Γ331 =
r
3
Γ
=
cot θ
=
32
Γ212 =
Γ233
Γ313
Γ323
Γ221 =
(348)
(349)
(350)
(351)
(352)
where the symbols are as explained above. As before, these results can be obtained from
Eqs. 334 and 336 using the scale factors of Table 1. As seen earlier (refer to § 2.2.2),
all these coordinate systems (i.e. orthonormal Cartesian, cylindrical and spherical) are
orthogonal systems, and hence Eqs. 334 and 336 do apply.
Because there is an element of arbitrariness in the choice of the order of coordinates,
and hence the order of their indices, the Christoffel symbols may be given in terms of
coordinate symbols rather than their indices to be more explicit and to avoid ambiguity
and confusion. For instance, in the above examples of the cylindrical coordinate systems
identified by the coordinates (ρ, φ, z) we may use [φφ, ρ] instead of [22, 1] and use Γφρφ
instead of Γ212 . Similarly, for the spherical coordinate systems identified by the coordinates
(r, θ, φ) we may use [θθ, r] instead of [22, 1] and use Γφrφ instead of Γ313 .
The Christoffel symbols of both kinds may also be subscripted or superscripted by the
gij
Γijk ) to reveal the
symbol of the metric tensor of the given space (e.g. [ij, k]g and
ij
metric which the symbols are based upon. This is especially important when we have
two or more different metrics related to two or more different spaces as it is the case, for
instance, in differential geometry of 2D surfaces embedded in a 3D space where we have
one metric for the surface and another metric for the 3D space. Alternatively, other means
of distinction may be used such as using Latin or upper case indices for the Christoffel
symbols of one metric and Greek or lower case indices for the Christoffel symbols of the
other metric, e.g. [ij, k] and Γijk for the Christoffel symbols of the space metric and [αβ, γ]
and Γγαβ for the Christoffel symbols of the surface metric. However, the latter methods do
not apply when the indices are numeric rather than symbolic.
The number of independent Christoffel symbols of each kind (first and second) in general
coordinate systems is given by:
n2 (n + 1)
(353)
NCI =
2
where n is the space dimension. The reason is that, due to the symmetry of the metric
independent metric components, gij , and for each independent
tensor there are n(n+1)
2
component there are n distinct Christoffel symbols. Alternatively, we have n3 permutations
of the three indices including the repetitive ones, and out of these n3 permutations we
have (n(n − 1)n) permutations whose paired indices are different where the three factors
correspond to the first, second and third index respectively. Now, due to the symmetry
of the Christoffel symbols in their paired indices, half of these (n(n − 1)n) permutations
are identical to the other half and hence they are not independent. Therefore, the total
5.2 Covariant Differentiation
120
number of independent Christoffel symbols is:
NCI = n3 −
2n3 − n3 + n2
n 3 + n2
n2 (n + 1)
(n (n − 1) n)
=
=
=
2
2
2
2
(354)
The partial derivative of the Christoffel symbol of the first kind is given by:
∂k [ij, l] =
1
(∂k ∂j gil + ∂k ∂i gjl − ∂k ∂l gij )
2
(355)
where this equation is based on the definition of the Christoffel symbol of the first kind as
given by Eq. 307.
Finally, we should remark that there are several notations for the Christoffel symbols.
As well as the symbols that we use in this book which may be the most common in
use (i.e. [ij, k] for the first kind and Γkij for the second kind), the first kind may also
while the second kind may be symbolized as kij or ij
be symbolized as Γijk or ij
k
k
as well as other notations. There may be some advantages or disadvantages in these
different notations. For example, the notations Γijk and Γkij may suggest, wrongly, that
these symbols are tensors which is not the case since the Christoffel symbols are not
general tensors although they are affine tensors. There may also be some advantages in
the typesetting and writing of these notations or there are factors related to recognition,
readability and even aesthetics.
5.2
Covariant Differentiation
The focus of this section is the operation of covariant differentiation of tensors which is a
generalization of the ordinary partial differentiation. The ordinary derivative of a tensor is
not a tensor in general. The objective of covariant differentiation is to ensure the invariance
of derivative (i.e. being a tensor) in general coordinate systems, and this results in applying
more sophisticated rules using Christoffel symbols where different differentiation rules for
covariant and contravariant indices apply. The resulting covariant derivative is a tensor
which is one rank higher than the differentiated tensor. In brief, the covariant derivative
is a partial derivative of the tensor that includes differentiating the basis vectors as well as
differentiating the components, as we will see. Hence, the covariant derivative of a general
tensor can be given generically by:
,m
i,··· ,m
j
n
j
n
∂k (Ai,···
j,··· ,n Ei · · · Em E · · · E ) = Aj,··· ,n;k Ei · · · Em E · · · E
(356)
,m
where the expression of Ai,···
j,··· ,n;k will be given in the following paragraphs (see e.g. Eq.
366).
More explicitly, the basis vectors in general curvilinear coordinate systems undergo
changes in magnitude and direction as they move around in their own space, and hence
they are functions of position. These changes should be accounted for when calculating the
derivatives of non-scalar tensors in such general systems. Therefore, terms based on using
Christoffel symbols are added to the ordinary derivative terms to correct for these changes
5.2 Covariant Differentiation
121
and this more comprehensive form of derivative is called the covariant derivative. Since
in rectilinear coordinate systems the basis vectors are constants, the Christoffel symbol
terms vanish identically and hence the covariant derivative reduces to the ordinary partial
derivative, but in the other coordinate systems these terms are present in general. As a
consequence, the ordinary derivative of a non-scalar tensor is a tensor iff the coordinate
transformations from Cartesian systems to that system are linear.
It has been suggested that the “covariant” label is an indication that the covariant differentiation operator ∇;i is in the covariant position. However, it may also be true that
“covariant” means “invariant” as the term “covariant” is also used in tensor calculus to
mean “invariant”. In fact, “covariant” as opposite to “contravariant” applies even to the
common form of ordinary partial differentiation since the commonly used partial differential operator ∂i is also in the covariant position. Anyway, contravariant differentiation,
which may be associated with the operator ∇;j , can also be defined for covariant and
contravariant tensors by raising the differentiation index of the covariant derivative using
the index raising operator, e.g.
Ai;j = g jk Ai;k
Ai;j = g jk Ai ;k
(357)
However, contravariant differentiation is rarely used.
As an example of how to obtain the covariant derivative of a tensor, let have a vector
A represented by contravariant components in general curvilinear coordinates, that is:
A = Ai Ei . We differentiate this vector following the normal rules of differentiation and
taking account of the fact that the basis vectors in general curvilinear coordinate systems
are differentiable functions of position and hence they, unlike their rectilinear counterparts,
are subject to differentiation using the product rule, that is:
(definition)
A;j = ∂j A
i
= ∂ j A Ei
= Ei ∂j Ai + Ai ∂j Ei
(product rule)
= Ei ∂j Ai + Ai Γkij Ek
(Eq. 312)
=
=
i
k
Ei ∂j A + A Γikj Ei
∂j Ai + Ak Γikj Ei
= Ai;j Ei
(358)
(relabeling dummy indices i ↔ k)
(taking common factor)
(notation)
where Ai;j , which is a rank-2 mixed tensor, is labeled the “covariant derivative” of Ai .
Similarly, for a vector represented by covariant components in general curvilinear coordinate systems, A = Ai Ei , we have:
(definition)
A;j = ∂j A
= ∂j Ai E
i
= Ei ∂j Ai + Ai ∂j Ei
(product rule)
= Ei ∂j Ai − Ai Γikj Ek
(Eq. 313)
(359)
5.2 Covariant Differentiation
122
= Ei ∂j Ai − Ak Γkij Ei
= ∂j Ai − Ak Γkij Ei
= Ai;j E
(relabeling dummy indices i ↔ k)
(taking common factor)
i
(notation)
The same rules apply to tensors of higher ranks. For example, for a rank-2 mixed tensor,
A = Ai j Ei Ej , we have:
(definition)
A;k = ∂k A
j
= ∂ k Ai E i E j
= ∂k Ai j Ei Ej
= ∂k Ai j Ei Ej
= ∂k Ai j Ei Ej
= ∂k Ai j Ei Ej
(product rule)
+ Ai j ∂k Ei Ej + Ai j Ei (∂k Ej )
j i
j
a
a
i
+ Ai −Γak E Ej + Ai E Γjk Ea (Eqs. 312 & 313)
− Ai j Γiak Ea Ej + Ai j Γajk Ei Ea
− Aaj Γaik Ei Ej + Ai a Γjak Ei Ej
= ∂k Ai j − Aaj Γaik + Ai a Γjak Ei Ej
= Ai
j
;k E
i
(360)
Ej
(Eq. 125)
(relabeling dummy indices)
(taking common factor)
(notation)
Based on the above arguments and examples, the main rules of covariant differentiation
can be outlined in the following examples. For a differentiable vector A, the covariant
derivative of the covariant and contravariant forms of the vector is given by:
Aj;i = ∂i Aj − Γkji Ak
Aj;i = ∂i Aj + Γjki Ak
(covariant)
(361)
(contravariant)
(362)
Similarly, for a differentiable rank-2 tensor A, the covariant derivative of the covariant,
contravariant and mixed forms of the tensor is given by:
Ajk;i = ∂i Ajk − Γlji Alk − Γlki Ajl
Ajk;i
Akj;i
jk
= ∂i A +
=
∂i Akj
+
Γjli Alk
Γkli Alj
+ Γkli Ajl
− Γlji Akl
(covariant)
(363)
(contravariant)
(364)
(mixed)
(365)
We note that for the mixed form there are two possibilities: one associated with the dyad
Ej Ek and the other with the dyad Ek Ej .
Following the methods and techniques outlined in the previous examples, we can easily
deduce the pattern of the operation of covariant differentiation. To obtain the covariant
derivative of a tensor in general, we start with an ordinary partial derivative term of the
component of the given tensor. Then for each tensor index an extra Christoffel symbol term
is added, positive for contravariant indices and negative for covariant indices, where the
differentiation index is one of the lower indices in the Christoffel symbol of the second kind.
Hence, for a differentiable rank-n tensor A in general coordinate systems the covariant
derivative with respect to the q th coordinate is given by:
ij...k
aj...k
ij...a
i
j
ia...k
k
Aij...k
lm...p;q = ∂q Alm...p + Γaq Alm...p + Γaq Alm...p + · · · + Γaq Alm...p
(366)
5.2 Covariant Differentiation
123
ij...k
ij...k
a
a
−Γalq Aij...k
am...p − Γmq Ala...p − · · · − Γpq Alm...a
In fact, there is practically only one possibility for the arrangement and labeling of the
indices in the Christoffel symbol terms of the covariant derivative of a tensor of any rank
if the following rules are observed:
1. The second subscript index of the Christoffel symbol is the differentiation index.
2. The differentiated index of the tensor in the Christoffel symbol term is contracted
with one of the indices of the Christoffel symbol using a new label and hence they
are opposite in their covariant and contravariant type.
3. The label of the differentiated index is transferred from the tensor to the Christoffel
symbol keeping its position as covariant or contravariant.
4. All the other indices of the tensor in the concerned Christoffel symbol term keep
their labels, position and order.
The ordinary partial derivative term in the covariant derivative expressions (see e.g. Eq.
366) represents the rate of change of the tensor components with change of position as a
result of moving along the coordinate curve of the differentiation index, while the Christoffel symbol terms represent the change experienced by the local basis vectors as a result of
the same movement. This can be seen from the development of Eqs. 358-360 where the
indicated terms correspond to the components and basis vectors according to the product
rule of differentiation.
From the above discussion, it is obvious that to obtain the covariant derivative, the
Christoffel symbols of the second kind should be obtained and these symbols are dependent
on the metric tensor. Hence, the covariant derivative is dependent on having the space
metric corresponding to the particular coordinate system. We also note that the covariant
derivative of a tensor is a tensor whose covariant rank is higher than the covariant rank
of the original tensor by one. Hence, the covariant derivative of a rank-n tensor of type
(r, s) is a rank-(n + 1) tensor of type (r, s + 1).
In all coordinate systems, the covariant derivative of a differentiable scalar function of
position, f , is the same as the ordinary partial derivative, that is:
f;i = f,i = ∂i f
(367)
This is justified by the fact that the covariant derivative is different from the ordinary
partial derivative because the basis vectors in general curvilinear coordinate systems are
dependent on their spatial position, and since a scalar is independent of the basis vectors
the covariant derivative and the partial derivative are identical. The derivation of the
expressions of the covariant derivative of contravariant and covariant vectors and higher
rank tensors, as given by Eqs. 358-360, clearly justifies this logic where the product rule
does not apply due to the absence of a basis vector in the representation of a scalar. This
can also be concluded from the covariant derivative rules as demonstrated in the previous
statements and formulated in the above equations like Eq. 366 since a scalar has no free
index and hence it cannot have any Christoffel symbol term. By a similar reasoning, since
the Christoffel symbols are identically zero in rectilinear coordinate systems, the covariant
derivative in these systems is the same as the ordinary partial derivative for all tensor
5.2 Covariant Differentiation
124
ranks, whether scalars or not. This can also be seen from the product rule (as employed in
the development of Eqs. 358-360 for instance) where all the terms involving differentiation
of basis vectors will vanish since these vectors are constant in rectilinear systems.
Another important fact about covariant differentiation is that the covariant derivative of
the metric tensor in its covariant, contravariant and mixed forms is zero in all coordinate
systems and hence it is treated like a constant with respect to covariant differentiation.
Accordingly, the covariant derivative operator bypasses the metric tensor, e.g.
(368)
∂;m gij Aj = gij ∂;m Aj
and hence the metric tensor commutes with the covariant differential operator. We will
expand on this issue later in this section.
Several rules of ordinary differentiation are naturally extended to covariant differentiation. For example, covariant differentiation is a linear operation with respect to algebraic
sums of tensor terms and hence the covariant derivative of a sum is the sum of the covariant
derivatives of the terms, that is:[22]
(aA ± bB);i = aA;i ± bB;i
(369)
where a and b are scalar constants and A and B are differentiable tensors. The product
rule of ordinary differentiation also applies to the covariant differentiation of inner and
outer products of tensors, that is:
(A ◦ B);i = A;i ◦ B + A ◦ B;i
(370)
where the symbol ◦ denotes an inner or outer product operator. However, as seen in this
equation, the order of the tensors should be observed since tensor multiplication, unlike
ordinary algebraic multiplication, is not commutative (refer to § 3.2.3). The product
rule is valid for the inner product of tensors because the inner product is an outer product
operation followed by a contraction of indices, and covariant differentiation and contraction
of indices do commute as we will see later.
A principal difference between partial differentiation and covariant differentiation is that
for successive differential operations with respect to different indices the ordinary partial
derivative operators do commute with each other, assuming that the well known continuity
condition is satisfied, but the covariant differential operators do not commute, that is:
∂i ∂j = ∂j ∂i
∂;i ∂;j 6= ∂;j ∂;i
(371)
This will be verified later in this section.
As indicated earlier, according to the “Ricci theorem” the covariant derivative of the
covariant, contravariant and mixed forms of the metric tensor is zero. This has nothing to
do with the metric tensor being a constant function of coordinates, which is true only for
[22]
We use a semicolon with symbolic notation of tensors in this equation and other similar equations for
the sake of clarity; the meaning should be obvious.
5.2 Covariant Differentiation
125
rectilinear coordinate systems, but this arises from the fact that the covariant derivative
quantifies the change with position of the basis vectors in magnitude and direction as well
as the change in components, and these contributions in the case of the metric tensor
cancel each other resulting in a total null effect. For example, for the covariant form of
the metric tensor we have:
g;k = ∂k gij Ei Ej
(definition)
(372)
j
i
i
i j
j
= (∂k gij ) E E + gij ∂k E E + gij E ∂k E
(product rule)
j
j
l
i
i
l
i j
(Eqs. 317 & 313)
= ([ik, j] + [jk, i]) E E + gij −Γlk E E + gij E −Γlk E
= [ik, j] Ei Ej + [jk, i] Ei Ej − gij Γilk El Ej − gij Γjlk Ei El
= [ik, j] Ei Ej + [jk, i] Ei Ej − [lk, j] El Ej − [lk, i] Ei El
i
j
i
j
i
j
i
= [ik, j] E E + [jk, i] E E − [ik, j] E E − [jk, i] E E
=0
j
(Eq. 309)
(relabeling dummy l)
Similarly, for the contravariant form of the metric tensor we have:
(definition)
(373)
g;k = ∂k g ij Ei Ej
(product rule)
= ∂k g ij Ei Ej + g ij (∂k Ei ) Ej + g ij Ei (∂k Ej )
j
(Eqs. 322 & 312)
= −g aj Γiak − g ia Γak Ei Ej + g ij (Γaik Ea ) Ej + g ij Ei Γajk Ea
= −g aj Γiak Ei Ej − g ia Γjak Ei Ej + g ij Γaik Ea Ej + g ij Γajk Ei Ea
= −g aj Γiak Ei Ej − g ia Γjak Ei Ej + g aj Γiak Ei Ej + g ia Γjak Ei Ej
=0
As for the mixed form of the metric tensor we have:
g;k = ∂k δ ij Ei Ej
= ∂k δ ij Ei Ej + δ ij (∂k Ei ) Ej + δ ij Ei ∂k Ej
= 0 + δ ij (Γaik Ea ) Ej + δ ij Ei −Γjak Ea
=
=
δ ij Γaik Ea Ej − δ ij Γjak Ei Ea
Γajk Ea Ej − Γiak Ei Ea
(relabeling indices)
(definition)
(374)
(product rule)
(Eqs. 312 & 313)
(Eq. 172)
= Γijk Ei Ej − Γijk Ei Ej
(relabeling dummy indices)
=0
As a result of the Ricci theorem, the metric tensor behaves as a constant with respect
to the covariant derivative operation, that is:
g;k = 0
(375)
where g is the metric tensor in its covariant or contravariant or mixed form, as seen above.
Consequently, the covariant derivative operator bypasses the metric tensor, that is:
(g ◦ A);k = g ◦ A;k
(376)
5.2 Covariant Differentiation
126
where A is a general tensor and the symbol ◦ denotes an inner or outer tensor product.[23]
The commutativity of the covariant derivative operator and the metric tensor acting as an
index shifting operator may be demonstrated more vividly by using the indicial notation,
e.g.
(377)
gik Ak;j = gik Ak;j = Ai;j = (Ai );j = gik Ak ;j
(378)
gmi gnj Amn;k = gmi gnj Amn;k = Aij;k = (Aij );k = (gmi gnj Amn );k
where we see that the sequence of these operations, seen in one order from the right and
in another order from the left, has no effect on the final result in the middle.
Like the metric tensor, the ordinary Kronecker delta tensor is constant with regard
to covariant differentiation and hence the covariant derivative of the Kronecker delta is
identically zero, that is:[24]
i
δj;k
= ∂k δji + δja Γiak − δai Γajk
= 0 + δja Γiak − δai Γajk
= 0 + Γijk − Γijk
(Eq. 365)
(379)
(δji is constant)
(Eq. 172)
=0
Accordingly, the covariant differential operator bypasses the Kronecker delta tensor which
is involved in inner and outer tensor products:[25]
(δ ◦ A);k = δ ◦ A;k
(380)
The rule of the Kronecker delta may be regarded as an instance of the rule of the metric
tensor, as stated by the Ricci theorem, since the Kronecker delta is a metric tensor for
certain systems and types. Like the ordinary Kronecker delta, the covariant derivative of
the generalized Kronecker delta is also identically zero. This may be deduced from Eq. 379
plus Eq. 199 where the generalized Kronecker delta is given as a determinant consisting of
ordinary Kronecker deltas and hence it is a sum of products of ordinary Kronecker deltas
whose partial derivative vanishes because each term in the derivative contains a derivative
of an ordinary Kronecker delta.
Based on the previous statements, we conclude that covariant differentiation and contraction of index operations commute with each other, e.g.
k
ij k
ij k
ik
ik
(381)
Aij
k;l δj = Ak;l δj = Ak;l = Ak ;l = Ak δj ;l
where we see again that the different sequences from the right and from the left produce
the same result in the middle. For clarity, we represented the contraction operation in this
[23]
Although the metric tensor is normally used in inner product operations for raising and lowering
indices, the possibility of its involvement in outer product operations should not be ruled out.
[24]
For diversity, we use indicial notation rather than symbolic notation which we used for example in
verifying g;k = 0.
[25]
Like the metric tensor, the Kronecker delta tensor is normally used in inner product operations for
replacement of indices; however the possibility of its involvement in outer product operations should
not be ruled out.
5.2 Covariant Differentiation
127
example by an inner product operation of the ordinary Kronecker delta tensor with the
contracted tensor.
For a differentiable function f (x, y) of class C 2 (i.e. all the second order partial derivatives
of the function do exist and are continuous), the mixed partial derivatives are equal, that
is:
∂x ∂y f = ∂y ∂x f
(382)
However, even if the components of a tensor satisfy this condition (i.e. being of class C 2 ),
this is not sufficient for the equality of the mixed covariant derivatives. What is required
for the mixed covariant derivatives to be equal is the vanishing of the Riemann-Christoffel
curvature tensor (see § 7.2.1) which is equivalent to having an intrinsically flat space. This
will be verified later in this section.
Higher order covariant derivatives are defined as derivatives of derivatives by successive
repetition of the process of covariant differentiation. However, the order of differentiation,
in the case of differentiating with respect to different indices, should be respected as stated
in the previous statements. For example, the second order mixed jk covariant derivative
of a contravariant vector A is given by:
(383)
Ai;jk = Ai;j ;k
= ∂k Ai;j + Γiak Aa;j − Γajk Ai;a
= ∂k ∂j Ai + Γiaj Aa + Γiak ∂j Aa + Γabj Ab − Γajk ∂a Ai + Γiba Ab
= ∂k ∂ j Ai + Γiaj ∂ k Aa + Aa ∂k Γiaj + Γiak ∂j Aa + Γiak Γabj Ab − Γajk ∂a Ai − Γajk Γiba Ab
that is:
Ai;jk = ∂k ∂ j Ai + Γiaj ∂ k Aa − Γajk ∂a Ai + Γiak ∂j Aa + Aa ∂k Γiaj − Γbjk Γiba + Γibk Γbaj
(384)
The second order mixed kj covariant derivative of a contravariant vector can be derived
similarly. However, it can be obtained more easily from the last equation by interchanging
the j and k indices, that is:
(385)
Ai;kj = ∂j ∂ k Ai + Γiak ∂ j Aa − Γakj ∂a Ai + Γiaj ∂k Aa + Aa ∂j Γiak − Γbkj Γiba + Γibj Γbak
The inequality of the jk and kj mixed derivatives in general can be verified by subtracting
the two sides of the last two equations from each other where the right hand side will not
vanish.
Similarly, the second order mixed jk covariant derivative of a covariant vector A is given
by:
Ai;jk = (Ai;j );k
= ∂k Ai;j − Γaik Aa;j − Γajk Ai;a
= ∂k ∂j Ai − Γbij Ab − Γaik ∂j Aa − Γbaj Ab − Γajk ∂a Ai − Γbia Ab
(386)
= ∂k ∂j Ai − Γbij ∂k Ab − Ab ∂k Γbij − Γaik ∂j Aa + Γaik Γbaj Ab − Γajk ∂a Ai + Γajk Γbia Ab
5.2 Covariant Differentiation
128
that is:
Ai;jk = ∂k ∂j Ai − Γaij ∂k Aa − Γaik ∂j Aa − Γajk ∂a Ai − Aa ∂k Γaij − Γbik Γabj − Γbjk Γaib
(387)
The second order mixed kj covariant derivative of a covariant vector can be obtained from
the last equation by interchanging the j and k indices, that is:
(388)
Ai;kj = ∂j ∂k Ai − Γaik ∂j Aa − Γaij ∂k Aa − Γakj ∂a Ai − Aa ∂j Γaik − Γbij Γabk − Γbkj Γaib
Again, the inequality of the jk and kj mixed derivatives can be verified by subtracting
the two sides of the last two equations from each other where the right hand side will not
vanish.
We will see in § 7.2.1 that the mixed second order covariant derivatives of a covariant vector Ai are linked through the Riemann-Christoffel curvature tensor Raijk by the following
relation:
Ai;jk − Ai;kj = Aa Raijk
(389)
This relation can be verified by subtracting the two sides of Eq. 388 from the two sides of
Eq. 387 and employing the definition of the Riemann-Christoffel curvature tensor of Eq.
560 to the right hand side, that is:
Ai;jk − Ai;kj = −Aa ∂k Γaij − Γbik Γabj − Γbjk Γaib + Aa ∂j Γaik − Γbij Γabk − Γbkj Γaib (390)
= Aa ∂j Γaik − Γbij Γabk − Γbkj Γaib − Aa ∂k Γaij − Γbik Γabj − Γbjk Γaib
= Aa ∂j Γaik − Γbij Γabk − Γbkj Γaib − ∂k Γaij + Γbik Γabj + Γbjk Γaib
= Aa ∂j Γaik − Γbij Γabk − ∂k Γaij + Γbik Γabj
= Aa Raijk
where the last step is based on Eq. 560.
The covariant derivatives of relative tensors, which are also relative tensors of the same
weight as the original tensors, are obtained by adding a weight term to the normal formulae
of covariant derivative. Hence, the covariant derivative of a relative scalar with weight w
is given by:
f;i = f,i − wf Γjji
(391)
while the covariant derivative of relative tensors of higher ranks with weight w is obtained
by adding the following term to the right hand side of Eq. 366:
a
− wAij...k
lm...p Γaq
(392)
Unlike ordinary differentiation, the covariant derivative of a non-scalar tensor with constant components is not zero in general curvilinear coordinate systems due to the presence
of the Christoffel symbols in the definition of the covariant derivative, as given by Eq. 366.
More explicitly, even though the partial derivative term is zero because the components
are constant, the Christoffel symbol terms are not zero in general because the basis vectors
are variables (refer to the derivation in Eqs. 358-360).
5.3 Absolute Differentiation
129
In rectilinear coordinate systems, the Christoffel symbols are identically zero because the
basis vectors are constants, and hence the covariant derivative is the same as the ordinary
partial derivative for all tensor ranks. As a result, when the components of the metric
tensor gij are constants, as it is the case for example in rectangular coordinate systems,
the covariant derivative becomes ordinary partial derivative. The constancy of the metric
tensor in this case can be concluded from the definition of the components of the metric
tensor as dot products of the basis vectors, as seen for example in Eqs. 213 and 214 (also
see Table 1).
We remark that for a differentiable covariant vector A which is a gradient of a scalar
field we have:
Ai;j = Aj;i
(393)
This can be easily verified by defining A as: Ai = f,i where f is a scalar and hence we
have:
(394)
Ai;j = (f,i );j
= (f,i ),j −
Γkij f,k
(Eq. 361)
= f,ij − Γkij f,k
= f,ji − Γkji f,k
(Eqs. 382 & 311)
(Eq. 361)
= (f,j );i
= Aj;i
Another important remark is that the covariant derivative of the basis vectors of the
covariant and contravariant types is identically zero, that is:
Ei;j = ∂j Ei − Γkij Ek = +Γkij Ek − Γkij Ek = 0
Ei;j = ∂j Ei + Γikj Ek = −Γikj Ek + Γikj Ek = 0
(395)
(396)
where Eqs. 361 and 362 are used in the first steps, while Eqs. 312 and 313 are used in the
second steps.
5.3
Absolute Differentiation
The absolute derivative of a tensor along a t-parameterized curve C(t) in an nD space
with respect to the parameter t is the inner product of the covariant derivative of the
tensor and the tangent vector to the curve. In brief, the absolute derivative is a covariant
derivative of a tensor along a curve. For a tensor Ai , the inner product of Ai;j , which is
i
a tensor, with another tensor is a tensor. Now, if the other tensor is du
, which is the
dt
tangent vector to a t-parameterized curve C(t) given by the equations ui = ui (t) where
i = 1, · · · , n, then the inner product:
Ai;r
dur
dt
(397)
5.3 Absolute Differentiation
130
is a tensor of the same rank and type as the tensor Ai . The tensor given by the expression
of Eq. 397 is called the “absolute” or “intrinsic” or “absolute covariant” derivative of the
i
, that is:
tensor Ai along the curve C and is symbolized by δA
δt
δAi
dur
= Ai;r
δt
dt
(398)
In fact, instead of basing the definition of absolute differentiation on the definition of the
covariant differentiation as a dot product of the covariant derivative with the tangent to
the curve as seen above, we can define absolute differentiation independently by following
the same method that we used to obtain the covariant derivative through applying the
differentiation process on the basis vectors as well as the components of the tensor (as seen
in the derivation of Eqs. 358-360) and hence the expressions of the absolute derivative
(e.g. the upcoming Eqs. 401 and 402) can be obtained directly by applying total differentiation to the tensor including its basis vectors. For example, the absolute derivative of
a contravariant vector A = Ai Ei can be obtained as follows:
d
dA
Ai Ei
=
dt
dt
dAi
dEi
+ Ai
= Ei
dt
dt
i
dA
∂Ei duj
= Ei
+ Ai j
dt
∂u dt
dAi
duj
= Ei
+ Ai Γkij Ek
dt
dt
i
duj
dA
+ Ei Ak Γikj
= Ei
dt
dti
j
dA
du
=
Ei
+ Ak Γikj
dt
dt
δAi
Ei
=
δt
(399)
(product rule)
(chain rule)
(Eq. 312)
(exchanging dummy indices i, k)
(definition of intrinsic derivative)
which is the same as the upcoming Eq. 401.
Following the above definitions, the absolute derivative of a differentiable scalar f is the
same as the ordinary total derivative, that is:
δf
df
=
δt
dt
(400)
This is because the covariant derivative of a scalar is the same as the ordinary partial
derivative or because a scalar has no association with basis vectors to differentiate. The
absolute derivative of a differentiable contravariant vector Ai with respect to the parameter
t is given by:
δAi
dAi
duj
=
+ Γikj Ak
(401)
δt
dt
dt
5.3 Absolute Differentiation
131
Similarly, for a differentiable covariant vector Ai we have:
dAi
duj
δAi
=
− Γkij Ak
δt
dt
dt
(402)
Absolute differentiation can be easily extended to higher rank (> 1) differentiable tensors
of type (m, n) along parameterized curves following the given examples and the pattern of
covariant derivative. For instance, the absolute derivative of a mixed tensor of type (1, 2)
Aijk along a t-parameterized curve C is given by:
dAijk
δAijk
dub
dub
dub
dub
= Aijk;b
=
+ Γiab Aajk
− Γajb Aiak
− Γabk Aija
δt
dt
dt
dt
dt
dt
(403)
Since the absolute derivative is given generically by:[26]
δA
duk
= A;k
δt
dt
(404)
it can be seen as an instance of the chain rule of differentiation where the two contracted
indices represent the in-between coordinate differential. This can also be concluded from
the above method of derivation of Eq. 399. Because the absolute derivative along a curve
is just an inner product of the covariant derivative with the tangent vector to the curve, the
well known rules of ordinary differentiation of sums and products also apply to absolute
differentiation, as for covariant differentiation, that is:
δ
δA
δB
(aA + bB) = a
+b
δt
δt
δt
δ
δB
δA
(A ◦ B) =
◦B + A◦
δt
δt
δt
(405)
(406)
where a and b are constant scalars, A and B are differentiable tensors and the symbol
◦ denotes an inner or outer product of tensors. However, the order of the tensors in the
products should be observed since tensor multiplication is not commutative.
Because absolute differentiation follows the style of covariant differentiation, the metric
tensor in its different variance types is in lieu of a constant with respect to absolute
differentiation, that is:
δgij
δg ij
=0
=0
(407)
δt
δt
and hence it passes through the absolute derivative operator, that is:
δAj
δ (gij Aj )
= gij
δt
δt
[26]
δ (g ij Aj )
δAj
= g ij
δt
δt
(408)
We use the absolute derivative notation with the symbolic notation of tensors in this equation and some
of the upcoming equations to ease the notation. This is similar to the use of the covariant derivative
notation with the symbolic notation as seen here and in previous equations. The meaning of these
notations should be clear
5.3 Absolute Differentiation
132
For coordinate systems in which all the components of the metric tensor are constants,
the absolute derivative is the same as the ordinary total derivative, as it is the case in
rectilinear coordinate systems, because the Christoffel symbol terms are zero in these
systems according to Eq. 308. The absolute derivative of a tensor along a given curve is
unique, and hence the ordinary derivative of the tensor along that curve in a rectangular
coordinate system is the same as the absolute derivative of the tensor along that curve in
any other system although the forms in the two systems may be different.
To sum up, we list in the following bullet points the main rules of covariant and absolute
differentiation:
1. Tensor differentiation (represented by covariant and absolute differentiation) is the
same as ordinary differentiation (represented by partial and total differentiation) but
with the application of the differentiation process not only on the tensor components
but also on the basis vectors that associate these components.
2. The sum and product rules of differentiation apply to covariant and absolute differentiation as for ordinary differentiation (i.e. partial and total).
3. The covariant and absolute derivatives of tensors are tensors.
4. The covariant and absolute derivatives of scalars and affine tensors of higher ranks
are the same as the ordinary derivatives.
5. The covariant and absolute derivative operators commute with the contraction of
indices.
6. The covariant and absolute derivatives of the metric, Kronecker and permutation
tensors as well as the basis vectors vanish identically in any coordinate system, that
is:
gij|q = 0
g ij|q = 0
(409)
i
δj|q
= 0
ij
= 0
δkl|q
(410)
ǫijk|q = 0
ǫijk
|q = 0
(411)
Ei|q = 0
Ei|q = 0
(412)
where the sign | represents covariant or absolute differentiation with respect to the
space coordinate uq . Hence, these tensors should be treated like constants in tensor
differentiation. This applies to the covariant and contravariant forms of these tensors,
as well as the mixed form when it is applicable. It also applies to the generalized, as
well as the ordinary, Kronecker delta as seen in the above equations.
7. Covariant differentiation increases the covariant rank of the differentiated tensor by
one, while absolute differentiation does not change the rank or type of the differentiated tensor.
8. The covariant and absolute derivatives in rectilinear systems are the same as the
partial and total derivatives respectively for all tensor ranks.
5.4 Exercises and Revision
5.4
133
Exercises and Revision
5.1 Why tensor differentiation (represented by covariant and absolute derivatives) is
needed in general coordinate systems to replace the ordinary differentiation (represented by partial and total derivatives)?
5.2 Show that in general coordinate systems, the ordinary differentiation of the components of non-scalar tensors with respect to the coordinates will not produce a tensor
in general.
5.3 “The Christoffel symbols are affine tensors but not tensors”. Explain and justify this
statement.
5.4 What is the difference between the first and second kinds of the Christoffel symbols?
5.5 Show that the Christoffel symbols of both kinds are not general tensors by giving
examples of these symbols being vanishing in some systems but not in other systems
and considering the universality of the zero tensor (see § 3.1.4).
5.6 State the mathematical definitions of the Christoffel symbols of the first and second
kinds. How these two kinds are transformed from each other?
5.7 What is the significance of the Christoffel symbols being solely dependent on the
coefficients of the metric tensor in their relation to the underlying space and coordinate
system?
5.8 Do the Christoffel symbols represent a property of the space, a property of the coordinate system, or a property of both?
5.9 If some of the Christoffel symbols vanish in a particular curvilinear coordinate system,
should these some necessarily vanish in other curvilinear coordinate systems? Justify
your answer by giving some examples.
5.10 Verify that the Christoffel symbols of the first and second kind are symmetric in their
paired indices by using their mathematical definitions.
5.11 Correct, if necessary, the following equations:
∂j Ei = −Γkij Ek
∂j Ei = −Γimj Em
5.12 What is the significance of the following equations?
Ek · ∂j Ei = Γkij
Ek · ∂j Ei = −Γikj
Ek · ∂j Ei = [ij, k]
5.13 Derive the following relations giving full explanation of each step:
√
∂j gil = [ij, l] + [lj, i]
Γjji = ∂i (ln g)
5.14 Assuming an orthogonal coordinate system, verify the following relation: [ij, k] = 0
where i 6= j 6= k.
5.15 Assuming an orthogonal coordinate system, verify the following relation: Γiji =
1
∂ ln gii with no sum over i.
2 j
5.16 Considering the identicality and difference of the indices of the Christoffel symbols of
either kind, how many cases we have? List these cases.
5.17 Prove the following relation which is used in Eq. 325: ∂i g = gg jl ∂i gjl .
5.4 Exercises and Revision
134
5.18 In orthogonal coordinate systems of a 3D space the number of independent nonidentically vanishing Christoffel symbols of either kind is only 15. Explain why.
5.19 Verify the following equations related to the Christoffel symbols in orthogonal coordinate systems in a 3D space:
[12, 1] = h1 h1,2
Γ323 =
h3,2
h3
5.20 Justify the following statement: “In any coordinate system, all the Christoffel symbols
of either kind vanish identically iff all the components of the metric tensor in the given
coordinate system are constants”.
5.21 Using Eq. 307 with Eq. 239, find the Christoffel symbols of the first kind corresponding to the Euclidean metric of cylindrical coordinate systems.
5.22 Give all the Christoffel symbols of the first and second kind of the following coordinate
systems: orthonormal Cartesian, cylindrical and spherical.
5.23 Mention two important properties of the Christoffel symbols of either kind with regard
to the order and similarity of their indices.
5.24 Using the entries in Eq. 336 and Table 1 and the properties of the Christoffel symbols
of the second kind, derive these symbols corresponding to the metrics of the coordinate
systems of question 5.22.
5.25 Write the following Christoffel symbols in terms of the coordinates instead of the
indices assuming a cylindrical system: [12, 1], [23, 1], Γ221 and Γ332 . Do the same
assuming a spherical system.
5.26 Show that all the Christoffel symbols will vanish when the components of the metric
tensor are constants.
5.27 Why the Christoffel symbols of either kind may be superscripted or subscripted by the
symbol of the underlying metric tensor? When this (or other measures for indicating
the underlying metric tensor) becomes necessary? Mention some of the other measures
used to indicate the underlying metric tensor.
5.28 Explain why the total number of independent Christoffel symbols of each kind is equal
2
to n (n+1)
.
2
5.29 Why covariant differentiation of tensors is regarded as a generalization of the ordinary
partial differentiation?
5.30 In general curvilinear coordinate systems, the variation of the basis vectors should
also be considered in the differentiation process of non-scalar tensors. Why?
5.31 State the mathematical definition of contravariant differentiation of a tensor Ai .
5.32 Obtain analytical expressions for Ai;j and B i;j by differentiating the vectors A = Ai Ei
and B = B i Ei .
5.33 Repeat question 5.32 with the rank-2 tensors C = Cij Ei Ej and D = Dij Ei Ej to
obtain Cij;k and Dij;k .
5.34 For a differentiable tensor A of type (m, n), the covariant derivative with respect to
the coordinate uk is given by:
i i ...i
m
Aji11ij22...i
...jn ;k =
∂Aj1 j2 ...jm
n
1 2
∂uk
im i1 i2 ...l
m
m
+ Γilk2 Aji11l...i
+Γilk1 Ajli12j...i
j2 ...jn + · · · + Γlk Aj1 j2 ...jn
2 ...jn
5.4 Exercises and Revision
135
i1 i2 ...im
m
2 ...im
− · · · − Γljn k Aji11ij22...i
− Γlj2 k Aij11il...j
−Γlj1 k Alj
...l
n
2 ...jn
5.35
5.36
5.37
5.38
5.39
5.40
5.41
5.42
5.43
5.44
5.45
5.46
5.47
5.48
5.49
5.50
5.51
5.52
5.53
Extract from the pattern of this expression the practical rules that should be followed
in writing the analytical expressions of covariant derivative of tensors of any rank and
type.
In the expression of covariant derivative, what the partial derivative term stands for
and what the Christoffel symbol terms represent?
For the covariant derivative of a type (m, n, w) tensor, obtain the number of total
terms, the number of negative Christoffel symbol terms and the number of positive
Christoffel symbol terms.
What is the rank and type of the covariant derivative of a tensor of rank-n and type
(p, q)?
The covariant derivative of a differentiable scalar function is the same as the ordinary
partial derivative. Why?
What is the significance of the dependence of the covariant derivative on the Christoffel
symbols with regard to its relation to the space and coordinate system?
The covariant derivative of tensors in coordinate systems with constant basis vectors
is the same as the ordinary partial derivative for all tensor ranks. Why?
Express, mathematically, the fact that the metric tensor is in lieu of constant with
respect to covariant differentiation.
Which rules of ordinary partial differentiation also apply to covariant differentiation
and which rules do not? State all these rules symbolically for both ordinary and
covariant differentiation.
Explain why the covariant differential operators with respect to different indices do
not commute, i.e. ∂;i ∂;j 6= ∂;j ∂;i (i 6= j).
State the Ricci theorem about covariant differentiation of the metric tensor and prove
it with full justification of each step.
State, symbolically, the commutative property of the covariant derivative operator
with the index shifting operator (which is based on the Ricci theorem) using the
symbolic notation one time and the indicial notation another.
Verify that the ordinary Kronecker delta tensor is constant with respect to covariant
differentiation.
State, symbolically, the fact that covariant differentiation and contraction of index
operations commute with each other.
What is the condition on the components of the metric tensor that makes the covariant
derivative become ordinary partial derivative for all tensor ranks?
Prove that covariant differentiation and contraction of indices commute.
What is the mathematical condition that is required if the mixed second order partial
derivatives should be equal, i.e. ∂i ∂j = ∂j ∂i (i 6= j)?
What is the mathematical condition that is required if the mixed second order covariant derivatives should be equal, i.e. ∂;i ∂;j = ∂;j ∂;i (i 6= j)?
Derive analytical expressions for Ai;jk and Ai;kj and hence verify that Ai;jk 6= Ai;kj .
From the result of exercise 5.52 plus Eq. 560, verify the following relation: Ai;jk −
5.4 Exercises and Revision
5.54
5.55
5.56
5.57
5.58
136
Ai;kj = Aa Raijk .
What is the covariant derivative of a relative scalar f of weight w? What is the
covariant derivative of a rank-2 relative tensor Aij of weight w?
Why the covariant derivative of a non-scalar tensor with constant components is not
necessarily zero in general coordinate systems? Which term of the covariant derivative
of such a tensor will vanish?
Show that: Ai;j = Aj;i where A is a gradient of a scalar field.
Show that the covariant derivative of the basis vectors of the covariant and contravariant types is identically zero, i.e. Ei;j = 0 and Ei;j = 0.
Prove the following identity:
∂k gij Ai B j = Ai;k B i + Ai Bi;k
5.59 Define absolute differentiation descriptively and mathematically. What are the other
names of absolute derivative?
5.60 Write the mathematical expression for the absolute derivative of the tensor field Aijk
which is defined over a space curve C(t).
5.61 Why the absolute derivative of a differentiable scalar is the same as its ordinary total
derivative, i.e. δf
= df
?
δt
dt
5.62 Why the absolute derivative of a differentiable non-scalar tensor is the same as its
ordinary total derivative in rectilinear coordinate systems?
5.63 From the pattern of covariant derivative of a general tensor, obtain the pattern of its
absolute derivative.
)
5.64 We have A = Aijk Ei Ej Ek . Apply the ordinary total differentiation process (i.e. dA
dt
onto this tensor (including its basis vectors) to obtain its absolute derivative.
5.65 Which rules of ordinary total differentiation also apply to intrinsic differentiation and
which rules do not? State all these rules symbolically for both ordinary and intrinsic
differentiation.
5.66 Using your knowledge about covariant differentiation and the fact that absolute differentiation follows the style of covariant differentiation, obtain all the rules of absolute
differentiation of the metric tensor, the Kronecker delta tensor and the index shifting
and index replacement operators. Express all these rules in words and in symbols.
5.67 Justify the following statement: “For coordinate systems in which all the components
of the metric tensor are constants, the absolute derivative is the same as the ordinary
total derivative”.
5.68 The absolute derivative of a tensor along a given curve is unique. What this means?
5.69 Summarize all the main properties and rules that govern covariant and absolute differentiation.
Chapter 6
Differential Operations
In this chapter, we examine the main differential operations which are based on the nabla
operator ∇ as defined in tensor calculus using largely tensor notation. These operations
are based on the various types of interaction between the vector differential operator
nabla and tensors of different ranks where some of these interactions involve the dot
and cross product operations. The chapter will investigate these operations in general
coordinate systems and in general orthogonal coordinate systems which are a special case
of the general coordinate systems. The chapter will also investigate these operations in
Cartesian coordinate systems as well as the two most important and widely used curvilinear
orthogonal coordinate systems, namely the cylindrical and spherical systems, because of
their particular importance and widespread application in science and engineering. Due
to the likely familiarity of the reader with these operations in Cartesian systems, which
is usually acquired at this level from a previous course on vector calculus, we start our
investigation from the Cartesian system. This may help to remind the reader and make
the subsequent sections easier to understand.
Regarding the cylindrical and spherical coordinate systems, we can use indexed general
coordinates like u1 , u2 and u3 to represent the cylindrical coordinates (ρ, φ, z) and the
spherical coordinates (r, θ, φ) and hence we can express these operations in tensor notation
as we do for the other systems. However, for the sake of clarity and to follow the more
conventional practice, we use the coordinates of these systems as suffixes in place of the
usual indices used in the tensor notation. In this context, we should insist that these
suffixes are labels and not indices and therefore they do not follow the rules and conventions
of tensor indices such as following the summation convention. In fact, there is another
reason for the use of suffix labels instead of symbolic indices that is the components in
the cylindrical and spherical coordinates are physical, not covariant or contravariant, and
hence suffixing with coordinates is more appropriate (see § 3.3).
Before we start this investigation, we should remark that the differentiation of a tensor
increases its rank by one, by introducing an extra covariant index, unless it implies a
contraction in which case it reduces the rank by one. Therefore the gradient of a scalar is
a vector and the gradient of a vector is a rank-2 tensor, while the divergence of a vector is
a scalar and the divergence of a rank-2 tensor is a vector. This may be justified by the fact
that the gradient operator is a vector operator. On the other hand the Laplacian operator
does not change the rank since it is a scalar operator; hence the Laplacian of a scalar is a
scalar and the Laplacian of a vector is a vector and so on.
We should also remark that there are other nabla based operators and operations which
are subsidiary to the main ones and are used in pure and applied mathematics and science.
For example, the following scalar differential operator, defined in Cartesian coordinates,
137
6.1 Cartesian Coordinate System
138
is commonly used in science such as fluid dynamics:
A · ∇ = Ai ∇i = Ai
∂
= Ai ∂i
∂xi
(413)
where A is a vector. As explained earlier (see § 1.2), the order of Ai and ∂i should be
respected. Another example is the following vector differential operator which also has
common applications in science:
[A × ∇]i = ǫijk Aj ∂k
(414)
where, again, the order should be respected and the operator is defined in Cartesian
coordinates.
6.1
6.1.1
Cartesian Coordinate System
Operators
The nabla vector operator ∇ is a spatial partial differential operator which is defined in
Cartesian coordinate systems by:
∂
∂xi
Similarly, the Laplacian scalar operator is given by:
∇i =
∇2 = δij
∂2
∂2
=
= ∇ii = ∂ii
∂xi ∂xj
∂xi ∂xi
(415)
(416)
We note that the Laplacian operator may also be notated with ∆ (as well as several other
symbols such as ∂ii2 ). However, in the present book we do not use this notation for this
purpose.
6.1.2
Gradient
Based on the above definition of nabla, the gradient of a differentiable scalar function of
position f is a vector obtained by applying the nabla operator to f and hence it is defined
by:
∂f
= ∂i f = f,i
(417)
[∇f ]i = ∇i f =
∂xi
Similarly, the gradient of a differentiable vector field A is the outer product (refer to §
3.2.3) between the nabla operator and the vector and hence it is a rank-2 tensor, that is:
[∇A]ij = ∂i Aj
This definition can be easily extended to higher rank tensors.
(418)
6.1.3 Divergence
6.1.3
139
Divergence
The divergence of a differentiable vector A is the dot product of the nabla operator and
the vector A and hence it is a scalar given by:
∇ · A = δij
∂Ai
∂Ai
=
= ∇i Ai = ∂i Ai = Ai,i
∂xj
∂xi
(419)
The divergence operation can also be viewed as taking the gradient of the vector followed
by a contraction. Hence, the divergence of a vector is invariant because it is the trace of
a rank-2 tensor (see § 7.1.2). It may also be argued more simply that the divergence of a
vector is a scalar and hence it is invariant.
Similarly, the divergence of a differentiable rank-2 tensor A is a vector defined in one of
its forms by:
[∇ · A]i = ∂j Aji
(420)
and in another form by:
(421)
[∇ · A]j = ∂i Aji
These two forms may be given respectively, using the symbolic notation, by:
∇·A
and
∇ · AT
(422)
where AT is the transpose of A. More generally, the divergence of a tensor of rank n ≥ 2,
which is a tensor of rank-(n − 1), can be defined in several forms, which are different in
general, depending on the choice of the contracted index.
6.1.4
Curl
The curl of a differentiable vector A is the cross product of the nabla operator and the
vector A and hence it is a vector defined by:
[∇ × A]i = ǫijk
∂Ak
= ǫijk ∇j Ak = ǫijk ∂j Ak = ǫijk Ak,j
∂xj
(423)
The curl operation may be generalized to tensors of rank > 1 (see for example § 7.1.4),
and hence the curl of a differentiable rank-2 tensor A can be defined as a rank-2 tensor
given by:
[∇ × A]ij = ǫimn ∂m Anj
(424)
The last example can be easily extended to higher rank tensors. We note that there is
more than one possibility for the contraction of the last index of the permutation tensor
with one of the tensor indices when the rank is > 1.
6.1.5 Laplacian
6.1.5
140
Laplacian
The Laplacian of a differentiable scalar f is given by:
∂ 2f
∂ 2f
=
= ∇ii f = ∂ii f = f,ii
∇ f = δij
∂xi ∂xj
∂xi ∂xi
2
(425)
The Laplacian of a differentiable vector A is defined for each component of the vector in
a similar manner to the definition of the Laplacian acting on a scalar, that is:
2
(426)
∇ A i = ∇2 [A]i = ∂jj Ai
This definition can be easily extended to higher rank tensors.
6.2
General Coordinate System
Here, we investigate the differential operators and operations in general coordinate systems. We note that the definitions of the differential operations in Cartesian systems, as
given in the vector calculus texts and as outlined in § 6.1, are essentially valid in general
non-Cartesian coordinate systems if the operations are extended to include the basis vectors as well as the components, as we will see in the following subsections. We also note
that the analytical expressions of the differential operations can be obtained directly if the
expression for the nabla operator ∇ and the spatial derivatives of the basis vectors of the
given coordinate system are known.
6.2.1
Operators
The nabla operator ∇, which is a spatial partial differential vector operator, is defined in
general coordinate systems by:
∇ = Ei ∂ i
(427)
Similarly, the Laplacian operator is defined generically by:
∇2 = div grad = ∇ · ∇
(428)
More details about these operators will follow in the next subsections.
6.2.2
Gradient
Based on the definition of the nabla operator, as given in the previous subsection, the
gradient of a differentiable scalar function of position, f , is given by:
∇f = Ei ∂i f = Ei f,i
(429)
The components of this expression represent the covariant form of a rank-1 tensor, i.e.
[∇f ]i = f,i , as it should be since the gradient operation increases the covariant rank of a
tensor by one, as indicated earlier. Since this expression consists of a contravariant basis
6.2.2 Gradient
141
vector and a covariant component, the gradient in general coordinate systems is invariant
under admissible coordinate transformations.
The gradient of a differentiable covariant vector A can similarly be defined as follows:[27]
(430)
∇A = Ei ∂i Aj Ej
= Ei Ej ∂i Aj + Ei Aj ∂i Ej
(product rule)
= Ei Ej ∂i Aj + Ei Aj −Γjki E
= Ei Ej ∂i Aj − Ei Ej Γkji Ak
= Ei Ej ∂i Aj − Γkji Ak
k
(Eq. 313)
(relabeling dummy indices j & k)
(taking common factor Ei Ej )
= Ei Ej Aj;i
(definition of covariant derivative)
and hence it is the covariant derivative of the vector. Similarly, for a differentiable contravariant vector A the gradient is given by:
(431)
∇A = Ei ∂i Aj Ej
= Ei Ej ∂i Aj + Ei Aj ∂i Ej
= Ei Ej ∂i Aj + Ei Aj Γkji Ek
=
=
=
Ei Ej ∂i Aj + Ei Ej Γjki Ak
Ei Ej ∂i Aj + Γjki Ak
Ei Ej Aj;i
(product rule)
(Eq. 312)
(relabeling dummy indices j & k)
(taking common factor Ei Ej )
(definition of covariant derivative)
The components of the gradients of covariant and contravariant vectors represent, respectively, the covariant and mixed forms of a rank-2 tensor, as they should be since the
gradient operation increases the covariant rank of a tensor by one.
The gradient of higher rank tensors is similarly defined. For example, the gradient of a
rank-2 tensor in its covariant, contravariant and mixed form is given by:
(covariant)
(432)
∇A = Ei Ej Ek ∂i Ajk − Γlji Alk − Γlki Ajl = Ei Ej Ek Ajk;i
jk
j
∇A = Ei Ej Ek ∂i Ajk + Γli Alk + Γkli Ajl = Ei Ej Ek A ;i
(contravariant) (433)
i j
k
k
l
k
k
l
i j
(mixed)
(434)
∇A = E E Ek ∂i Aj − Γji Al + Γli Aj = E E Ek Aj;i
j
j l
j
i
k j
l
i
k
(mixed)
(435)
∇A = E Ej E ∂i A k + Γli A k − Γki A l = E Ej E A k;i
We finally remark that the contravariant form of the gradient operation can be obtained
by using the index raising operator. For example, the contravariant form of the gradient
of a scalar f is given by:
[∇f ]i = ∂ i f = g ij ∂j f = g ij f,j = f ,i
[27]
(436)
We note that the basis vector Ei which associates the derivative operator in the following equations
(as well as in similar equations and expressions) should be the last one in the basis tensor so that the
order of the indices inthe components and in the basis tensor are the same. So, strictly we should
write ∇A = ∂i Aj Ej Ei where ∂i acts only on what is inside the parentheses and hence the final
expression becomes ∇A = Ej Ei Aj;i . However, to avoid confusion we put the vector to the left relying
on this understanding.
6.2.3 Divergence
142
This can be easily extended to higher rank tensors. As seen, the contravariant form of the
gradient operation increases the contravariant rank of the tensor by one.
6.2.3
Divergence
Generically, the divergence of a differentiable contravariant vector A is defined as follows:
∇ · A = E i ∂ i · Aj E j
(437)
= Ei · ∂i Aj Ej
(definition of covariant derivative)
= Ei · Aj;i Ej
j
= Ei · Ej A ;i
= δ ij Aj;i
(Eq. 215)
Ai;i
(Eq. 172)
=
In more details, the divergence of a differentiable contravariant vector Ai is a scalar obtained by contracting the covariant derivative index with the contravariant index of the
vector, and hence:
∇ · A = Ai;i
(438)
i
= ∂i A +
Γiji Aj
1
√
= ∂i Ai + Aj √ ∂j ( g)
g
1
√
= ∂i Ai + Ai √ ∂i ( g)
g
1
√ i
= √ ∂i
gA
g
(Eq. 362)
(Eq. 324)
(renaming dummy index j)
(product rule)
where g is the determinant of the covariant metric tensor gij . The last equality may be
called the Voss-Weyl formula.
The divergence can also be obtained by raising the first index of the covariant derivative
of a covariant vector using a contracting contravariant metric tensor, that is:
(Eq. 376)
(439)
g ji Aj;i = g ji Aj ;i
(Eq. 228)
= Ai ;i
= Ai;i
=∇·A
(Eq. 438)
as before. Accordingly, the divergence of a covariant vector Aj is obtained by using the
raising operator, that is:
Ai;i = g ij Aj;i
(440)
6.2.4 Curl
143
For a rank-2 contravariant tensor A, the divergence is generically defined by:
∇ · A = Ei ∂i · Ajk Ej Ek
(441)
= Ei · ∂i Ajk Ej Ek
jk
i
(definition of covariant derivative)
= E · A ;i Ej Ek
= Ei · Ej Ek Ajk;i
= δji Ek Ajk;i
(Eq. 215)
= Ek Aik;i
(Eq. 172)
The components of this expression represent a contravariant vector, as it should be since
the divergence operation reduces the contravariant rank of a tensor by one.
More generally, considering the tensor components, the divergence of a differentiable
rank-2 contravariant tensor Aij is a contravariant vector obtained by contracting the covariant derivative index with one of the contravariant indices, that is:
[∇ · A]j = Aij;i
or
[∇ · A]i = Aij;j
(442)
And for a rank-2 mixed tensor Aij we have:
[∇ · A]j = Aij;i
(443)
k ···im
Similarly, for a general tensor of type (m, n): A = Aji11ij22···i
·········jn , the divergence with respect
to its k th contravariant index is defined by:
···im
i1 i2 ···s···im
=
A
[∇ · A]ji11ij22············j
(444)
j1 j2 ·········jn ;s
n
with the absence of the contracted contravariant index ik on the left hand side. As a
matter of notation, it should be understood that ∇ · A is lower than the original tensor
A by just one contravariant index and hence, unlike the common use of this notation, it
is not necessarily scalar.
6.2.4
Curl
The curl of a differentiable vector is the cross product of the nabla operator ∇ with the
vector. For example, the curl of a vector A represented by covariant components is given
by:
(445)
curl A = ∇ × A
= Ei ∂i × Aj Ej
= Ei × ∂i Aj Ej
= Ei × Aj;i Ej
= Aj;i Ei × Ej
(Eq. 359)
6.2.5 Laplacian
144
= Aj;i ǫijk Ek
(Eq. 274)
ijk
= ǫ Aj;i Ek
ǫijk
= √ ∂i Aj − Γlji Al Ek
g
(Eqs. 166 & 361)
Hence, the k th contravariant component of curl A is given by:
ǫijk
[∇ × A]k = √ ∂i Aj − Γlji Al
g
(446)
On expanding the last equation for the three components of a vector in a 3D space,
considering that the terms of the Christoffel symbols cancel out due to their symmetry in
the two lower indices,[28] we obtain:
1
[∇ × A]1 = √ (∂2 A3 − ∂3 A2 )
g
1
[∇ × A]2 = √ (∂3 A1 − ∂1 A3 )
g
1
[∇ × A]3 = √ (∂1 A2 − ∂2 A1 )
g
(447)
(448)
(449)
Hence, Eq. 446 will reduce to:
ǫijk
[∇ × A]k = √ ∂i Aj
g
6.2.5
(450)
Laplacian
Generically, the Laplacian of a differentiable scalar function of position, f , is defined as
follows:
∇2 f = div (grad f ) = ∇ · (∇f )
(451)
Hence the simplest approach for obtaining the Laplacian in general coordinate systems
is to insert the expression for the gradient, ∇f , into the expression for the divergence.
However, because in general coordinate systems the divergence is defined only for tensors
having at least one contravariant index to facilitate the contraction with the covariant
[28]
That is:
Ai;j − Aj;i
=
=
=
=
∂j Ai − Ak Γkij − ∂i Aj + Ak Γkji
∂j Ai − Ak Γkij − ∂i Aj + Ak Γkij
∂j Ai − ∂i Aj
Ai,j − Aj,i
6.2.5 Laplacian
145
derivative index (see for example Eqs. 437 and 444) whereas the gradient of a scalar is
a covariant tensor, the index of the gradient should be raised first before applying the
divergence operation, that is:
[∇f ]i = ∂ i f = g ij ∂j f
(452)
Now, according to Eq. 438 the divergence is given by:
1
√ i
∇ · A = Ai;i = √ ∂i
gA
g
(453)
On defining A ≡ Ei ∂ i f and replacing Ai in Eq. 453 with ∂ i f using Eq. 452 we obtain:
1
√ ij
∇2 f = ∇ · Ei ∂ i f = √ ∂ i
gg ∂j f
g
(454)
which is the expression for the Laplacian of a scalar function f in general coordinate
systems.
Another approach for developing the Laplacian expression in general coordinate systems
is to apply the first principles by using the definitions and basic properties of the operations
involved, that is:
∇2 f = ∇ · (∇f )
i
(455)
j
= E ∂i · E ∂j f
= Ei · ∂ i Ej ∂ j f
= Ei · ∂i Ej f,j
= Ei · Ej f,j;i
= Ei · Ej f,j;i
(Eq. 359)
= g ij f,j;i
ij
= g f,j
= g ij ∂j f
ij
(Eq. 214)
;i
(Eq. 376)
(Eq. 362)
;i
= ∂i g ∂j f + g kj ∂j f Γiki
= ∂i g ij ∂j f + g ij ∂j f Γkik
1
√
= ∂i g ij ∂j f + g ij ∂j f √ (∂i g)
g
√
1 √
=√
g∂i g ij ∂j f + g ij ∂j f ∂i g
g
1
√ ij
= √ ∂i
gg ∂j f
g
(renaming dummy indices i & k)
(Eq. 324)
1
(taking √ factor)
g
(product rule)
which is the same as before (see Eq. 454).
The Laplacian of a scalar f may also be shorthand notated with:
∇2 f = g ij f,ij
(456)
6.3 Orthogonal Coordinate System
146
The Laplacian of non-scalar tensors can be similarly defined. For example, the Laplacian
of a vector B in its contravariant and covariant forms, B i and Bi , is a vector A (i.e.
A = ∇2 B) which may be defined in general coordinate systems as:
Ai = g jk B i ;jk
Ai = g jk Bi;jk
(457)
As indicated earlier, the Laplacian of a tensor is a tensor of the same rank and variance
type.
6.3
Orthogonal Coordinate System
In this section, we state the main differential operators and operations in general orthogonal coordinate systems. These operators and operations are special cases of those of
general coordinate systems which were derived in § 6.2. However, due to the widespread
use of orthogonal coordinate systems, it is beneficial to state the most important of these
operators and operations although they can be easily obtained from the formulae of general
coordinate systems.
For clarity, general orthogonal coordinate systems are identified in this section by the
coordinates (q 1 , . . . , q n ) with unit basis vectors (q1 , . . . , qn ) and scale factors (h1 , . . . , hn )
where:
X 1 ∂xj
X ∂q i
qi =
e
=
h i j ej
(no sum on i)
(458)
j
hi ∂q i
∂x
j
j
#1/2 "
#−1/2
"
X ∂q i 2
X ∂xj 2
∂r
=
hi = |Ei | =
=
∂q i
∂q i
∂xj
j
j
(459)
In the last equations, xj and ej are respectively the coordinates and unit basis vectors in
a Cartesian rectangular system, and r = xi ei is the position vector in that system. We
remark that in orthogonal coordinate systems the covariant and contravariant normalized
basis vectors are identical, as established previously in § 3.3, and hence qi = qi and
ej = ej .
6.3.1
Operators
The nabla operator in general orthogonal coordinate systems is given by:
∇=
X qi ∂
hi ∂q i
i
(460)
while the Laplacian operator, assuming a 3D space, is given by:
3
1 X ∂
h1 h2 h3 ∂
∇ =
h1 h2 h3 i=1 ∂q i
(hi )2 ∂q i
2
(461)
6.3.2 Gradient
6.3.2
147
Gradient
The gradient of a differentiable scalar f in orthogonal coordinate systems, assuming a 3D
space, is given by:
3
X
q1 ∂f
q2 ∂f
q3 ∂f
qi ∂f
=
+
+
∇f =
i
1
2
h ∂q
h1 ∂q
h2 ∂q
h3 ∂q 3
i=1 i
6.3.3
(462)
Divergence
The divergence in orthogonal coordinate systems can be obtained from Eq. 438. Since
for orthogonal coordinate systems the metric tensor according to Eqs. 233-236 is diagonal
√
with g = h1 h2 h3 in a 3D space and hi Ai = Âi (no summation) according to Eq. 144,
the last line of Eq. 438 becomes:
1 ∂ √ i
gA
∇·A = √
g ∂q i
3
1 X ∂
h1 h2 h3 i
=
Â
h1 h2 h3 i=1 ∂q i
hi
∂
∂
1
∂
h2 h3 Â1 + 2 h1 h3 Â2 + 3 h1 h2 Â3
=
h1 h2 h3 ∂q 1
∂q
∂q
(463)
where A is a contravariant differentiable vector and Âi represents its physical components.
This equation is the divergence of a vector in general orthogonal coordinates as defined in
vector calculus. We note that in orthogonal coordinate systems the physical components
are the same for covariant and contravariant forms, as established before in § 3.3, and
hence Âi = Âi .
6.3.4
Curl
The curl of a differentiable vector A in orthogonal coordinate systems in 3D spaces is
given by:
h1 q 1 h2 q2 h3 q3
1
∂
∂
∂
∇×A=
(464)
∂q 1
∂q 2
∂q 3
h1 h2 h3
h1 Â1 h2 Â2 h3 Â3
where the hat indicates a physical component. The last equation may also be given in a
more compact form as:
3
X
ǫijk hi ∂(hk Âk )
[∇ × A]i =
h1 h2 h3 ∂q j
k=1
(no sum on i)
(465)
6.3.5 Laplacian
6.3.5
148
Laplacian
As seen earlier, for general orthogonal coordinate systems in 3D spaces we have:
√
g ii =
g = h1 h2 h3
1
(hi )2
(no sum)
g ij = 0 (i 6= j)
(466)
and hence Eq. 454 becomes:
3
1 X ∂
h1 h2 h3 ∂f
∇f=
h1 h2 h3 i=1 ∂q i
(hi )2 ∂q i
2
(467)
which is the Laplacian of a scalar function of position, f , in orthogonal coordinate systems
as defined in vector calculus.
6.4
Cylindrical Coordinate System
For cylindrical coordinate systems identified by the coordinates (ρ, φ, z), the orthonormal
basis vectors are eρ , eφ and ez . Although the given components (i.e. Aρ , Aφ and Az ) are
physical components, we do not use hats since the components are suffixed with coordinate
symbols (refer to § 3.3). We use for brevity eρφ as a shorthand notation for the unit dyad
eρ eφ and similar notations for the other dyads. We remark that the following expressions
for the operators and operations in cylindrical coordinate systems can be obtained from the
definition of these operators and operations in general coordinate systems using the metric
tensor of the cylindrical system (see § 4.5). They can also be obtained more simply from
the corresponding expressions in orthogonal coordinate systems using the scale factors of
the cylindrical system in Table 1.
It should be obvious that since ρ, φ and z are labels for specific coordinates and not
variable indices, the summation convention does not apply to these labels. We note that
cylindrical coordinate systems are defined specifically for 3D spaces. We also note that
the following operators and operations can be obtained for the 2D plane polar coordinate
systems by dropping the z components or terms from the cylindrical form of these operators
and operations.
6.4.1
Operators
The nabla operator ∇ in cylindrical coordinate systems is given by:
1
∇ = eρ ∂ ρ + eφ ∂ φ + ez ∂ z
ρ
(468)
while the Laplacian operator is given by:
1
1
∇2 = ∂ρρ + ∂ρ + 2 ∂φφ + ∂zz
ρ
ρ
where ∂ρρ = ∂ρ ∂ρ and the notation equally applies to other similar symbols.
(469)
6.4.2 Gradient
6.4.2
149
Gradient
The gradient of a differentiable scalar f is given by:
1
∇f = eρ ∂ρ f + eφ ∂φ f + ez ∂z f
ρ
(470)
The gradient of a differentiable vector A is given by:
∇A = eρρ Aρ,ρ + eρφ Aφ,ρ + eρz Az,ρ +
1
1
1
Aφ
Aρ
+ eφφ
+ eφz Az,φ +
Aρ,φ −
Aφ,φ +
eφρ
ρ
ρ
ρ
ρ
ρ
ezρ Aρ,z + ezφ Aφ,z + ezz Az,z
(471)
where eρρ , eρφ , · · · , ezz are unit dyads as defined above.
6.4.3
Divergence
The divergence of a differentiable vector A is given by:
∇·A=
1
[∂ρ (ρAρ ) + ∂φ Aφ + ρ∂z Az ]
ρ
The divergence of a differentiable rank-2 tensor A is given by:
Aρρ − Aφφ 1
+ Aφρ,φ + Azρ,z +
∇ · A = eρ Aρρ,ρ +
ρ
ρ
2Aρφ 1
Aφρ − Aρφ
eφ Aρφ,ρ +
+ Aφφ,φ + Azφ,z +
+
ρ
ρ
ρ
Aρz 1
ez Aρz,ρ +
+ Aφz,φ + Azz,z
ρ
ρ
(472)
(473)
We note that Eq. 472 can be obtained by contracting Eq. 471.
6.4.4
Curl
The curl of a differentiable vector A is given by:
∇×A=
6.4.5
1
ρ
eρ ρeφ ez
∂ρ ∂φ ∂z
Aρ ρAφ Az
(474)
Laplacian
The Laplacian of a differentiable scalar f is given by:
1
1
∇2 f = ∂ρρ f + ∂ρ f + 2 ∂φφ f + ∂zz f
ρ
ρ
(475)
6.5 Spherical Coordinate System
The Laplacian of a differentiable vector A is given by:
2
1
1
2
∂ρ (ρAρ ) + 2 ∂φφ Aρ + ∂zz Aρ − 2 ∂φ Aφ +
∇ A = eρ ∂ ρ
ρ
ρ
ρ
1
2
1
eφ ∂ ρ
∂ρ (ρAφ ) + 2 ∂φφ Aφ + ∂zz Aφ + 2 ∂φ Aρ +
ρ
ρ
ρ
1
1
ez ∂ρ (ρ∂ρ Az ) + 2 ∂φφ Az + ∂zz Az
ρ
ρ
6.5
150
(476)
Spherical Coordinate System
For spherical coordinate systems identified by the coordinates (r, θ, φ), the orthonormal
basis vectors are er , eθ and eφ . As in the case of cylindrical coordinates, the components
are physical and we do not use hats for the same reason. We use for brevity erθ as
a shorthand notation for the dyad er eθ and similar notations for the other unit dyads.
As for cylindrical systems, the following expressions for the operators and operations in
spherical coordinate systems can be obtained from the corresponding definitions in general
coordinate systems using the metric tensor of the spherical system (see § 4.5) or from the
corresponding expressions in orthogonal systems using the scale factors of the spherical
system in Table 1. Again, the summation convention does not apply to r, θ and φ since
they are labels and not indices. We also note that spherical coordinate systems are defined
specifically for 3D spaces.
6.5.1
Operators
The nabla operator ∇ in spherical coordinate systems is given by:
1
1
∂φ
∇ = er ∂ r + eθ ∂ θ + eφ
r
r sin θ
(477)
while the Laplacian operator is given by:
cos θ
1
1
2
∂θ + 2 2 ∂φφ
∇2 = ∂rr + ∂r + 2 ∂θθ + 2
r
r
r sin θ
r sin θ
(478)
where ∂rr = ∂r ∂r and the notation equally applies to other similar symbols.
6.5.2
Gradient
The gradient of a differentiable scalar f in spherical coordinate systems is given by:
1
1
∇f = er ∂r f + eθ ∂θ f + eφ
∂φ f
r
r sin θ
(479)
The gradient of a differentiable vector A is given by:
∇A = err Ar,r + erθ Aθ,r + erφ Aφ,r +
(480)
6.5.3 Divergence
151
Ar,θ Aθ
Aθ,θ Ar
Aφ,θ
−
+ eθθ
+
+ eθφ
+
eθr
r
r
r
r
r
Aθ,φ
Aφ,φ
Aφ
Aφ cot θ
Ar Aθ cot θ
Ar,φ
+ eφθ
+ eφφ
−
−
+
+
eφr
r sin θ
r
r sin θ
r
r sin θ
r
r
where err , erθ , · · · , eφφ are unit dyads as defined above.
6.5.3
Divergence
The divergence of a differentiable vector A is given by:
∂ (r2 Ar )
∂ (sin θAθ )
∂Aφ
1
sin θ
+r
+r
∇·A= 2
r sin θ
∂r
∂θ
∂φ
(481)
The divergence of a differentiable rank-2 tensor A is given by:
∂r (r2 Arr ) ∂θ (Aθr sin θ) ∂φ Aφr Aθθ + Aφφ
∇ · A = er
+
+
−
+
(482)
r2
r sin θ
r sin θ
r
∂r (r3 Arθ ) ∂θ (Aθθ sin θ) ∂φ Aφθ Aθr − Arθ − Aφφ cot θ
eθ
+
+
+
+
r3
r sin θ
r sin θ
r
∂r (r3 Arφ ) ∂θ (Aθφ sin θ) ∂φ Aφφ Aφr − Arφ + Aφθ cot θ
+
eφ
+
+
r3
r sin θ
r sin θ
r
We note that Eq. 481 can be obtained by contracting Eq. 480.
6.5.4
Curl
The curl of a differentiable vector A is given by:
∇×A=
6.5.5
1
r2 sin θ
er reθ r sin θeφ
∂r ∂θ
∂φ
Ar rAθ r sin θAφ
(483)
Laplacian
The Laplacian of a differentiable scalar f is given by:
2
cos θ
1
1
∇2 f = ∂rr f + ∂r f + 2 ∂θθ f + 2
∂θ f + 2 2 ∂φφ f
(484)
r
r
r sin θ
r sin θ
The Laplacian of a differentiable vector A is given by:
2∂θ (Aθ sin θ)
∂θ (sin θ∂θ Ar )
∂φφ Ar
2∂φ Aφ
∂r (r2 Ar )
2
+
+ 2 2 −
− 2
+(485)
∇ A = er ∂ r
r2
r2 sin θ
r2 sin θ
r sin θ
r sin θ
∂r (r2 ∂r Aθ )
1
∂θ (Aθ sin θ)
2 cot θ
∂φφ Aθ
2∂θ Ar
eθ
+ 2 2 +
+ 2 ∂θ
− 2
∂ φ Aφ +
r2
r
sin θ
r2
r sin θ
r sin θ
∂φφ Aφ
1
∂θ (Aφ sin θ)
2∂φ Ar
2 cot θ
∂r (r2 ∂r Aφ )
+ 2 2 + 2
+ 2 ∂θ
+
∂ φ Aθ
eφ
r2
r
sin θ
r sin θ r sin θ r2 sin θ
6.6 Exercises and Revision
6.6
152
Exercises and Revision
6.1 Describe briefly the nabla based differential operators and operations considering the
interaction of the nabla operator with the tensors which are acted upon by this operator.
6.2 What are the advantages and disadvantages of using the coordinates as suffixes for
labeling the operators, basis vectors and tensor components in cylindrical and spherical systems instead of indexed general coordinates? What are the advantages and
disadvantages of the opposite?
6.3 “The differentiation of a tensor increases its rank by one, by introducing an extra
covariant index, unless it implies a contraction in which case it reduces the rank by
one”. Justify this statement giving common examples from vector and tensor calculus.
6.4 Write the following subsidiary nabla based operators in tensor notation: A · ∇ and
A × ∇. Is this notation consistent with the notation of dot and cross product of
vectors?
6.5 Why in general we have: A · ∇ =
6 ∇ · A and A × ∇ 6= ∇ × A?
6.6 Define the nabla vector operator and the Laplacian scalar operator in Cartesian coordinate systems using tensor notation.
6.7 Find the gradient of the following vector field in a Cartesian coordinate system: A =
(x, 2x2 , π).
6.8 Define the divergence of a differentiable vector descriptively and mathematically assuming a Cartesian coordinate system.
6.9 What is the divergence of the following vector field in Cartesian coordinates: A =
(2z, y 3 , ex )?
6.10 Write symbolically, using tensor notation, the following two forms of the divergence
of a rank-2 tensor field A in Cartesian coordinates: ∇ · A and ∇ · AT .
6.11 Define the curl ∇ × A in Cartesian coordinates using tensor notation where (a) A
is a rank-1 tensor and (b) A is a rank-2 tensor (note the two possibilities in the last
case).
6.12 What is the curl of the following vector field assuming a Cartesian coordinate system:
A = (5e2x , πxy, z 2 )?
6.13 Find the Laplacian of the following vector field in Cartesian coordinates: A =
(x2 y, 2y sin z, πzecosh x ).
6.14 Define the nabla operator and the Laplacian operator in general coordinate systems
using tensor notation.
6.15 Obtain an expression for the gradient of a covariant vector A = Ai Ei in general coordinates justifying each step in your derivation. Repeat the question with a contravariant
vector A = Ai Ei .
6.16 Repeat question 6.15 with a rank-2 mixed tensor A = Aij Ei Ej .
6.17 Define, in tensor language, the contravariant form of the gradient of a scalar field.
6.18 Define the divergence of a differentiable vector descriptively and mathematically assuming a general coordinate system.
6.19 Derive the following expression for the divergence of a contravariant vector A in
6.6 Exercises and Revision
6.20
6.21
6.22
6.23
153
general coordinates: ∇ · A = Ai;i .
Verify the following formula for the
divergence of a contravariant vector A in general
√ i
coordinates: ∇ · A = √1g ∂i
gA . Repeat the question with the formula: ∇ · A =
g ji Aj;i where A is a covariant vector.
Repeat question 6.20 with the formula: ∇ · A = Ek Aik;i where A is a rank-2 contravariant tensor.
Prove that the divergence of a contravariant vector is a scalar (i.e. rank-0 tensor) by
showing that it is invariant under coordinate transformations.
Derive, from the first principles, the following formula for the curl
of a covariant vector
ijk
field A in general coordinates: [∇ × A]k = ǫ√g ∂i Aj − Γlji Al .
6.24 Show that the formula in exercise 6.23 will reduce to [∇ × A]k = ǫ√g ∂i Aj due to the
symmetry of the Christoffel symbols in their lower indices.
6.25 Derive, from the first principles, the following expression
for the Laplacian of a scalar
√ ij
field f in general coordinates: ∇2 f = √1g ∂i
gg ∂j f .
6.26 Why the basic definition of the Laplacian of a scalar field f in general coordinates
as ∇2 f = div (grad f ) cannot be used as it is to develop a formula before raising the
index of the gradient?
6.27 Define, in tensor language, the nabla operator and the Laplacian operator assuming
an orthogonal coordinate system of a 3D space.
6.28 Using the expression of the divergence of a vector field in general coordinates, obtain
an expression for the divergence in orthogonal coordinates.
6.29 Define the curl of a vector field A in orthogonal coordinates of a 3D space using
determinantal form and tensor notation form.
6.30 Using the expression of the Laplacian of a scalar field in general coordinates, derive
an expression for the Laplacian in orthogonal coordinates of a 3D space.
6.31 Why the components of tensors in cylindrical and spherical coordinates are physical?
6.32 Define the nabla and Laplacian operators in cylindrical coordinates.
6.33 Use the definition of the gradient of a scalar field f in orthogonal coordinates and
the scale factors of Table 1 to obtain an expression for the gradient in cylindrical
coordinates.
6.34 Use the definition of the divergence of a vector field A in orthogonal coordinates and
the scale factors of Table 1 to obtain an expression for the divergence in cylindrical
coordinates.
6.35 Write the determinantal form of the curl of a vector field A in cylindrical coordinates.
6.36 Use the definition of the Laplacian of a scalar field f in orthogonal coordinates and
the scale factors of Table 1 to obtain an expression for the Laplacian in cylindrical
coordinates.
6.37 A scalar field in cylindrical coordinates is given by: f (ρ, φ, z) = ρ. What are the
gradient and Laplacian of this field?
6.38 A vector field in cylindrical coordinates is given by: A (ρ, φ, z) = (3z, πφ2 , z 2 cos ρ).
What are the divergence and curl of this field?
6.39 Repeat exercise 6.33 with spherical coordinates.
ijk
6.6 Exercises and Revision
6.40
6.41
6.42
6.43
154
Repeat exercise 6.34 with spherical coordinates.
Repeat exercise 6.35 with spherical coordinates.
Repeat exercise 6.36 with spherical coordinates.
A scalar field in spherical coordinates is given by: f (r, θ, φ) = r2 + θ. What are the
gradient and Laplacian of this field?
6.44 A vector field in spherical coordinates is given by: A (r, θ, φ) = (er , 5 sin φ, ln θ). What
are the divergence and curl of this field?
Chapter 7
Tensors in Application
In this chapter, we conduct a preliminary investigation about some tensors and tensor
notation and techniques which are commonly used in the mathematical and physical applications of tensor calculus. The chapter is made of three sections dedicated to tensors in
mathematics, geometry and science. The mathematics part (see § 7.1) comes from tensor
applications in linear algebra and vector calculus, while most of the materials in the science part (see § 7.3) come from applications related to fluid and continuum mechanics. In
the geometry part (see § 7.2) a few prominent tensors and tensor identities of wide applications in differential geometry and related scientific fields are examined. We should also
refer the reader to the previous chapter (see § 6) as an example of tensor applications in
mathematics and science since the materials in that chapter are partly based on employing
tensor notation and techniques.
We note that all the aforementioned disciplines, where the materials about tensors come
from, are intimately linked to tensor calculus since large parts of this subject were developed within those disciplines. Also, the materials presented in this chapter about tensors
and tensor techniques are used as vital building blocks and tools in a number of important
mathematical and physical theories. However, we would like to insist that although these
materials provide a very useful glimpse, they are just partially representative examples of
tensor applications. Our objective is to have more familiarity with some prominent tensors
and tensor techniques and hence they are not meant to provide a comprehensive view. We
should also indicate that some tensors, especially those in the science section, are defined
differently in different disciplines and hence the given definitions and descriptions may not
be thorough or universal.
7.1
Tensors in Mathematics
In this section, we provide a sample of common definitions related to basic concepts and
operations in matrix and vector algebra using the language of tensor calculus. Common
identities in vector calculus as well as the integral theorems of Gauss and Stokes and
some important scalar invariants are also examined in this section from this perspective.
Finally, we provide a rather extensive set of examples about the use of tensor language
and techniques in proving mathematical identities where these identities are gathered
from vector calculus. For simplicity, clarity and wide applicability we employ a Cartesian
approach in the tensor formulations of this section.
155
7.1.1 Common Definitions in Tensor Notation
7.1.1
156
Common Definitions in Tensor Notation
The trace of a matrix A representing a rank-2 tensor in an nD space is given by:
tr (A) = Aii
(i = 1, . . . , n)
(486)
For a 3 × 3 matrix representing a rank-2 tensor in a 3D space, the determinant is given
by:
A11 A12 A13
det (A) = A21 A22 A23 = ǫijk A1i A2j A3k = ǫijk Ai1 Aj2 Ak3
(487)
A31 A32 A33
where the last two equalities represent the expansion of the determinant by row and by
column. Alternatively, the determinant of a 3 × 3 matrix can be given by:
det (A) =
1
ǫijk ǫlmn Ail Ajm Akn
3!
(488)
More generally, for an n × n matrix representing a rank-2 tensor in an nD space, the
determinant is given by:
det (A) = ǫi1 ···in A1i1 . . . Anin
= ǫi1 ···in Ai1 1 . . . Ain n
1
=
ǫi ···i ǫj ···j Ai j . . . Ain jn
n! 1 n 1 n 1 1
(489)
The inverse of a matrix A representing a rank-2 tensor is given by:
A−1
ij
=
1
ǫipq ǫjmn Amp Anq
2 det (A)
(490)
The multiplication of a matrix A by a vector b, as defined in linear algebra, is given by:
[Ab]i = Aij bj
(491)
It should be remarked that we are using matrix notation in the writing of Ab. According
to the symbolic notation of tensors, the multiplication operation should be denoted by a
dot between the symbols of the tensor and the vector, i.e. A·b, since matrix multiplication
in matrix algebra is equivalent to an inner product operation in tensor algebra.
Similarly, the multiplication of two compatible matrices A and B, as defined in linear
algebra, is given by:
[AB]ik = Aij Bjk
(492)
Again, we are using here matrix notation in the writing of AB; otherwise a dot should be
inserted between the symbols of the two matrices.
The dot product of two vectors of the same dimension is given by:
A · B = δij Ai Bj = Ai Bi
(493)
7.1.2 Scalar Invariants of Tensors
157
Similarly, the cross product of two vectors in a 3D space is given by:
[A × B]i = ǫijk Aj Bk
(494)
The scalar triple product of three vectors in a 3D space is given by:
A · (B × C) =
A1 A2 A3
B1 B2 B3
C1 C2 C3
= ǫijk Ai Bj Ck
(495)
while the vector triple product of three vectors in a 3D space is given by:
[A × (B × C)]i = ǫijk ǫklm Aj Bl Cm
(496)
The expression of the other principal form of the vector triple product, i.e. (A × B) × C,
can be obtained from the above form by changing the order of the factors in the external
cross product and reversing the sign. Other operations, like relabeling the indices and
exchanging some of the indices of the epsilons with a shift in sign, can then follow to
obtain a more organized form.
7.1.2
Scalar Invariants of Tensors
In the following, we list and write in tensor notation a number of invariants of low rank
tensors which have special importance due to their widespread applications in vector and
tensor calculus. All these invariants are scalars.
The value of a scalar (rank-0 tensor), which consists of a magnitude and a sign, is
invariant under coordinate transformations. An invariant of a vector (rank-1 tensor) under
coordinate transformations is its magnitude, i.e. length.[29] The main three independent
scalar invariants of a rank-2 tensor A are:
I = tr (A) = Aii
II = tr A2 = Aij Aji
III = tr A3 = Aij Ajk Aki
(497)
(498)
(499)
I1 = I = Aii
1
1 2
I2 =
I − II = (Aii Ajj − Aij Aji )
2
2
1
1 3
I − 3I II + 2III = ǫijk ǫpqr Aip Ajq Akr
I3 = det (A) =
3!
3!
(500)
Different forms of the three invariants of a rank-2 tensor A, which are also widely used,
are the following (noting that some of these definitions may belong to 3D specifically):
[29]
(501)
(502)
The direction is also invariant but it is not a scalar! In fact the magnitude alone is invariant under
coordinate transformations even for pseudo vectors because it is a true scalar.
7.1.3 Common Identities in Vector and Tensor Notation
158
where I, II and III are given by Eqs. 497-499. As the invariants I1 , I2 and I3 are defined
in terms of the invariants I, II and III according to Eqs. 500-502, the invariants I, II
and III can also be defined in terms of the invariants I1 , I2 and I3 , that is:
I = I1
II = I12 − 2I2
III = I13 − 3I1 I2 + 3I3
(503)
(504)
(505)
Since the determinant of a matrix representing a rank-2 tensor is invariant (see Eq. 502),
then if the determinant vanishes in one coordinate system, it will vanish in all transformed
coordinate systems, and if not it will not (also refer to § 3.1.4). Consequently, if a rank-2
tensor is invertible in a particular coordinate system, it will be invertible in all coordinate
systems, and if not it will not.
The following are ten common scalar invariants that are jointly formed between two rank2 tensors A and B: tr (A), tr (B), tr (A2 ), tr (B2 ), tr (A3 ), tr (B3 ), tr (A · B), tr (A2 · B),
tr (A · B2 ) and tr (A2 · B2 ). As seen, all these are traces.
7.1.3
Common Identities in Vector and Tensor Notation
In this subsection, we present some of the widely used identities of vector calculus using
the traditional vector notation as well as its equivalent tensor notation. In the following
equations, f and h are differentiable scalar fields; A, B, C and D are differentiable vector
fields; and r = xi ei is the position vector.
∇·r = n
∂ i xi = n
(vector notation)
(tensor notation)
(506)
(507)
where n is the space dimension.
∇×r = 0
ǫijk ∂j xk = 0
(508)
(509)
∇ (a · r) = a
∂i (aj xj ) = ai
(510)
(511)
∇ · (∇f ) = ∇2 f
∂i (∂i f ) = ∂ii f
(512)
(513)
∇ · (∇ × A) = 0
ǫijk ∂i ∂j Ak = 0
(514)
(515)
where a is a constant vector.
7.1.3 Common Identities in Vector and Tensor Notation
159
∇ × (∇f ) = 0
ǫijk ∂j ∂k f = 0
(516)
(517)
∇ (f h) = f ∇h + h∇f
∂i (f h) = f ∂i h + h∂i f
(518)
(519)
∇ · (f A) = f ∇ · A + A · ∇f
∂i (f Ai ) = f ∂i Ai + Ai ∂i f
(520)
(521)
∇ × (f A) = f ∇ × A + ∇f × A
ǫijk ∂j (f Ak ) = f ǫijk ∂j Ak + ǫijk (∂j f ) Ak
(522)
(523)
A · (B × C) = C · (A × B) = B · (C × A)
ǫijk Ai Bj Ck = ǫkij Ck Ai Bj = ǫjki Bj Ck Ai
(524)
(525)
A × (B × C) = B (A · C) − C (A · B)
ǫijk Aj ǫklm Bl Cm = Bi (Am Cm ) − Ci (Al Bl )
(526)
(527)
A × (∇ × B) = (∇B) · A − A · ∇B
ǫijk ǫklm Aj ∂l Bm = (∂i Bm ) Am − Al (∂l Bi )
(528)
(529)
∇ × (∇ × A) = ∇ (∇ · A) − ∇2 A
ǫijk ǫklm ∂j ∂l Am = ∂i (∂m Am ) − ∂ll Ai
(530)
(531)
∇ (A · B) = A × (∇ × B) + B × (∇ × A) + (A · ∇) B + (B · ∇) A
(532)
∂i (Am Bm ) = ǫijk Aj (ǫklm ∂l Bm ) + ǫijk Bj (ǫklm ∂l Am ) + (Al ∂l ) Bi + (Bl ∂l ) Ai (533)
∇ · (A × B) = B · (∇ × A) − A · (∇ × B)
∂i (ǫijk Aj Bk ) = Bk (ǫkij ∂i Aj ) − Aj (ǫjik ∂i Bk )
(534)
(535)
∇ × (A × B) = (B · ∇) A + (∇ · B) A − (∇ · A) B − (A · ∇) B
ǫijk ǫklm ∂j (Al Bm ) = (Bm ∂m ) Ai + (∂m Bm ) Ai − (∂j Aj ) Bi − (Aj ∂j ) Bi
(536)
(537)
A·C A·D
B·C B·D
= (Al Cl ) (Bm Dm ) − (Am Dm ) (Bl Cl )
(A × B) · (C × D) =
ǫijk Aj Bk ǫilm Cl Dm
(538)
(539)
7.1.4 Integral Theorems in Tensor Notation
160
(A × B) × (C × D) = [D · (A × B)] C − [C · (A × B)] D
ǫijk ǫjmn Am Bn ǫkpq Cp Dq = (ǫqmn Dq Am Bn ) Ci − (ǫpmn Cp Am Bn ) Di
(540)
(541)
In vector and tensor notations, the condition for a vector field A to be solenoidal is given
by:
(542)
(543)
∇·A = 0
∂i Ai = 0
In vector and tensor notations, the condition for a vector field A to be irrotational is given
by:
(544)
(545)
∇×A = 0
ǫijk ∂j Ak = 0
7.1.4
Integral Theorems in Tensor Notation
The divergence theorem for a differentiable vector field A in vector and tensor notations
is given by:
¨
˚
A · n dσ
(546)
∇ · A dτ =
S
Ωˆ
ˆ
Ai ni dσ
(547)
∂i Ai dτ =
S
Ω
where Ω is a bounded region in an nD space enclosed by a generalized surface S, dτ and
dσ are generalized volume and area differentials, n and ni are the unit vector normal to
the surface and its ith component, and the index i ranges over 1, . . . , n.
Similarly, the divergence theorem for a differentiable rank-2 tensor field A in tensor
notation for the first index is given by:
ˆ
ˆ
∂i Ail dτ =
Ail ni dσ
(548)
Ω
S
while the divergence theorem for differentiable tensor fields of higher rank A in tensor
notation for the index k is given by:
ˆ
ˆ
∂k Aij...k...m dτ =
Aij...k...m nk dσ
(549)
Ω
S
Stokes theorem for a differentiable vector field A in vector and tensor notations is given
by:
¨
ˆ
(∇ × A) · n dσ =
A · dr
(550)
S
C
ˆ
ˆ
Ai dxi
(551)
ǫijk ∂j Ak ni dσ =
S
C
7.1.5 Examples of Using Tensor Techniques to Prove Identities
161
where C stands for the perimeter of the surface S, and dr is a differential of the position
vector which is tangent to the perimeter while the other symbols are as defined above.
Similarly, Stokes theorem for a differentiable rank-2 tensor field A in tensor notation for
the first index is given by:
ˆ
ˆ
Ail dxi
(552)
ǫijk ∂j Akl ni dσ =
S
C
while Stokes theorem for differentiable tensor fields of higher rank A in tensor notation
for the index k is given by:
ˆ
ˆ
Alm...k...n dxk
(553)
ǫijk ∂j Alm...k...n ni dσ =
S
7.1.5
C
Examples of Using Tensor Techniques to Prove Identities
In this subsection, we provide some examples for using tensor techniques to prove vector and tensor identities where comments are added next to each step to explain and
justify. These examples, which are based on the identities given in § 7.1.3, demonstrate
the elegance, efficiency and clarity of the methods and notation of tensor calculus. We
note that in Cartesian coordinate systems, some tensor equations in general coordinates
change their form and hence in the following we give the corresponding Cartesian form of
the equations which are not given previously in this form since we use the Cartesian form
in these proofs. The added comments inside the parentheses refer to the corresponding
equations in general coordinates:
ǫijk ǫlmk
δij Aj
∂xi
∂xj
∂ i xi
• ∇ · r = n:
• ∇ × r = 0:
= δil δjm − δim δjl
= Ai
(Eq. 167)
(Eq. 172)
(554)
(555)
= ∂j xi = xi,j = δij
(Eq. 175)
(556)
= δii = n
(Eq. 176)
(557)
∇ · r = ∂ i xi
= δii
=n
[∇ × r]i = ǫijk ∂j xk
= ǫijk δkj
= ǫijj
=0
(Eq. 419)
(Eq. 557)
(Eq. 557)
(Eq.
(Eq.
(Eq.
(Eq.
Since i is a free index, the identity is proved for all components.
423)
556)
555)
155)
7.1.5 Examples of Using Tensor Techniques to Prove Identities
• ∇ (a · r) = a:
[∇ (a · r)]i = ∂i (aj xj )
= aj ∂i xj + xj ∂i aj
= aj ∂i xj
= aj δji
= ai
= [a]i
162
(Eqs. 417 & 493)
(product rule)
(aj is constant)
(Eq. 556)
(Eq. 555)
(definition of index)
Since i is a free index, the identity is proved for all components.
• ∇ · (∇f ) = ∇2 f :
∇ · (∇f ) = ∂i [∇f ]i
(Eq. 419)
= ∂i (∂i f )
(Eq. 417)
= ∂i ∂i f
(rules of differentiation)
= ∂ii f
(definition of 2nd derivative)
2
=∇ f
(Eq. 425)
• ∇ · (∇ × A) = 0:
∇ · (∇ × A) = ∂i [∇ × A]i
= ∂i (ǫijk ∂j Ak )
= ǫijk ∂i ∂j Ak
= ǫijk ∂j ∂i Ak
= −ǫjik ∂j ∂i Ak
= −ǫijk ∂i ∂j Ak
=0
(Eq. 419)
(Eq. 423)
(ǫ is constant)
(continuity condition)
(Eq. 180)
(relabeling dummy indices i and j)
(since ǫijk ∂i ∂j Ak = −ǫijk ∂i ∂j Ak )
This can also be concluded from line three by arguing that: since by the continuity condition ∂i and ∂j can change their order with no change in the value of the term while a
corresponding change of the order of i and j in ǫijk results in a sign change, we see that
each term in the sum has its own negative and hence the terms add up to zero.
• ∇ × (∇f ) = 0:
[∇ × (∇f )]i = ǫijk ∂j [∇f ]k
(Eq. 423)
= ǫijk ∂j (∂k f )
(Eq. 417)
= ǫijk ∂j ∂k f
(rules of differentiation)
= ǫijk ∂k ∂j f
(continuity condition)
= −ǫikj ∂k ∂j f
(Eq. 180)
= −ǫijk ∂j ∂k f
(relabeling dummy indices j and k)
=0
(since ǫijk ∂j ∂k f = −ǫijk ∂j ∂k f )
This can also be concluded from line three by a similar argument to the one given in the
previous point. Because [∇ × (∇f )]i is an arbitrary component, then each component is
zero.
7.1.5 Examples of Using Tensor Techniques to Prove Identities
• ∇ (f h) = f ∇h + h∇f :
[∇ (f h)]i = ∂i (f h)
= f ∂i h + h∂i f
= [f ∇h]i + [h∇f ]i
= [f ∇h + h∇f ]i
163
(Eq. 417)
(product rule)
(Eq. 417)
(Eq. 24)
Because i is a free index, the identity is proved for all components.
• ∇ · (f A) = f ∇ · A + A · ∇f :
∇ · (f A) = ∂i [f A]i
(Eq. 419)
= ∂i (f Ai )
(definition of index)
= f ∂i Ai + Ai ∂i f
(product rule)
= f ∇ · A + A · ∇f
(Eqs. 419 & 413)
• ∇ × (f A) = f ∇ × A + ∇f × A:
[∇ × (f A)]i = ǫijk ∂j [f A]k
= ǫijk ∂j (f Ak )
= f ǫijk ∂j Ak + ǫijk (∂j f ) Ak
= f ǫijk ∂j Ak + ǫijk [∇f ]j Ak
= [f ∇ × A]i + [∇f × A]i
= [f ∇ × A + ∇f × A]i
(Eq. 423)
(definition of index)
(product rule & commutativity)
(Eq. 417)
(Eqs. 423 & 494)
(Eq. 24)
Because i is a free index, the identity is proved for all components.
• A · (B × C) = C · (A × B) = B · (C × A):
A · (B × C) = ǫijk Ai Bj Ck
(Eq. 495)
= ǫkij Ai Bj Ck
(Eq. 180)
= ǫkij Ck Ai Bj
(commutativity)
= C · (A × B)
(Eq. 495)
= ǫjki Ai Bj Ck
(Eq. 180)
= ǫjki Bj Ck Ai
(commutativity)
= B · (C × A)
(Eq. 495)
The negative permutations of this identity can be similarly obtained and proved by changing the order of the vectors in the cross products which results in a sign change.
• A × (B × C) = B (A · C) − C (A · B):
[A × (B × C)]i = ǫijk Aj [B × C]k
(Eq. 494)
= ǫijk Aj ǫklm Bl Cm
(Eq. 494)
= ǫijk ǫklm Aj Bl Cm
(commutativity)
= ǫijk ǫlmk Aj Bl Cm
(Eq. 180)
= (δil δjm − δim δjl ) Aj Bl Cm
(Eq. 554)
7.1.5 Examples of Using Tensor Techniques to Prove Identities
= δil δjm Aj Bl Cm − δim δjl Aj Bl Cm
= (δil Bl ) (δjm Aj Cm ) − (δim Cm ) (δjl Aj Bl )
= Bi (Am Cm ) − Ci (Al Bl )
= Bi (A · C) − Ci (A · B)
= [B (A · C)]i − [C (A · B)]i
= [B (A · C) − C (A · B)]i
164
(distributivity)
(commutativity & grouping)
(Eq. 555)
(Eq. 493)
(definition of index)
(Eq. 24)
Because i is a free index, the identity is proved for all components. The other variant
of this identity, i.e. (A × B) × C, can be obtained and proved similarly by changing the
order of the factors in the external cross product with adding a minus sign.
• A × (∇ × B) = (∇B) · A − A · ∇B:
[A × (∇ × B)]i = ǫijk Aj [∇ × B]k
(Eq. 494)
= ǫijk Aj ǫklm ∂l Bm
(Eq. 423)
= ǫijk ǫklm Aj ∂l Bm
(commutativity)
= ǫijk ǫlmk Aj ∂l Bm
(Eq. 180)
= (δil δjm − δim δjl ) Aj ∂l Bm
(Eq. 554)
= δil δjm Aj ∂l Bm − δim δjl Aj ∂l Bm
(distributivity)
= Am ∂i Bm − Al ∂l Bi
(Eq. 555)
= (∂i Bm ) Am − Al (∂l Bi )
(commutativity & grouping)
= [(∇B) · A]i − [A · ∇B]i
(Eq. 418 & § 3.2.5)
= [(∇B) · A − A · ∇B]i
(Eq. 24)
Because i is a free index, the identity is proved for
• ∇ × (∇ × A) = ∇ (∇ · A) − ∇2 A:
[∇ × (∇ × A)]i = ǫijk ∂j [∇ × A]k
= ǫijk ∂j (ǫklm ∂l Am )
= ǫijk ǫklm ∂j (∂l Am )
= ǫijk ǫlmk ∂j ∂l Am
= (δil δjm − δim δjl ) ∂j ∂l Am
= δil δjm ∂j ∂l Am − δim δjl ∂j ∂l Am
= ∂m ∂i Am − ∂l ∂l Ai
= ∂i (∂m Am ) − ∂ll Ai
= [∇ (∇ · A)]i − ∇2 A i
= ∇ (∇ · A) − ∇2 A i
all components.
(Eq. 423)
(Eq. 423)
(ǫ is constant)
(Eq. 180 & definition of derivative)
(Eq. 554)
(distributivity)
(Eq. 555)
(∂ shift, grouping & Eq. 2)
(Eqs. 419, 417 & 426)
(Eq. 24)
Because i is a free index, the identity is proved for all components. This identity can
also be considered as an instance of the identity before the last one, observing that in the
second term on the right hand side the Laplacian should precede the vector, and hence no
independent proof is required.
7.1.5 Examples of Using Tensor Techniques to Prove Identities
• ∇ (A · B) = A × (∇ × B) + B × (∇ × A) + (A · ∇) B + (B · ∇) A:
We start from the right hand side and end with the left hand side:
[A × (∇ × B) + B × (∇ × A) + (A · ∇) B + (B · ∇) A]i =
[A × (∇ × B)]i + [B × (∇ × A)]i + [(A · ∇) B]i + [(B · ∇) A]i =
(Eq. 24)
ǫijk Aj [∇ × B]k + ǫijk Bj [∇ × A]k + (Al ∂l ) Bi + (Bl ∂l ) Ai =
(Eqs. 494, 419 & indexing)
ǫijk Aj (ǫklm ∂l Bm ) + ǫijk Bj (ǫklm ∂l Am ) + (Al ∂l ) Bi + (Bl ∂l ) Ai =
(Eq. 423)
ǫijk ǫklm Aj ∂l Bm + ǫijk ǫklm Bj ∂l Am + (Al ∂l ) Bi + (Bl ∂l ) Ai =
(commutativity)
ǫijk ǫlmk Aj ∂l Bm + ǫijk ǫlmk Bj ∂l Am + (Al ∂l ) Bi + (Bl ∂l ) Ai =
(Eq. 180)
(δil δjm − δim δjl ) Aj ∂l Bm + (δil δjm − δim δjl ) Bj ∂l Am + (Al ∂l ) Bi + (Bl ∂l ) Ai =
(Eq. 554)
(δil δjm Aj ∂l Bm − δim δjl Aj ∂l Bm ) + (δil δjm Bj ∂l Am − δim δjl Bj ∂l Am )
+ (Al ∂l ) Bi + (Bl ∂l ) Ai =
(distributivity)
δil δjm Aj ∂l Bm − Al ∂l Bi + δil δjm Bj ∂l Am − Bl ∂l Ai + (Al ∂l ) Bi + (Bl ∂l ) Ai =
(Eq. 555)
δil δjm Aj ∂l Bm − (Al ∂l ) Bi + δil δjm Bj ∂l Am − (Bl ∂l ) Ai + (Al ∂l ) Bi + (Bl ∂l ) Ai =
(grouping)
δil δjm Aj ∂l Bm + δil δjm Bj ∂l Am =
(cancellation)
Am ∂i Bm + Bm ∂i Am =
(Eq. 555)
∂i (Am Bm ) =
(product rule)
[∇ (A · B)]i
(Eqs. 417 & 419)
Because i is a free index, the identity is proved for all components.
• ∇ · (A × B) = B · (∇ × A) − A · (∇ × B):
∇ · (A × B) = ∂i [A × B]i
(Eq. 419)
= ∂i (ǫijk Aj Bk )
(Eq. 494)
= ǫijk ∂i (Aj Bk )
(ǫ is constant)
= ǫijk (Bk ∂i Aj + Aj ∂i Bk )
(product rule)
165
7.1.5 Examples of Using Tensor Techniques to Prove Identities
= ǫijk Bk ∂i Aj + ǫijk Aj ∂i Bk
= ǫkij Bk ∂i Aj − ǫjik Aj ∂i Bk
= Bk (ǫkij ∂i Aj ) − Aj (ǫjik ∂i Bk )
= Bk [∇ × A]k − Aj [∇ × B]j
= B · (∇ × A) − A · (∇ × B)
166
(distributivity)
(Eq. 180)
(commutativity & grouping)
(Eq. 423)
(Eq. 493)
• ∇ × (A × B) = (B · ∇) A + (∇ · B) A − (∇ · A) B − (A · ∇) B:
[∇ × (A × B)]i = ǫijk ∂j [A × B]k
= ǫijk ∂j (ǫklm Al Bm )
= ǫijk ǫklm ∂j (Al Bm )
= ǫijk ǫklm (Bm ∂j Al + Al ∂j Bm )
= ǫijk ǫlmk (Bm ∂j Al + Al ∂j Bm )
= (δil δjm − δim δjl ) (Bm ∂j Al + Al ∂j Bm )
= δil δjm Bm ∂j Al + δil δjm Al ∂j Bm −
δim δjl Bm ∂j Al − δim δjl Al ∂j Bm
= Bm ∂m Ai + Ai ∂m Bm − Bi ∂j Aj − Aj ∂j Bi
= (Bm ∂m ) Ai + (∂m Bm ) Ai − (∂j Aj ) Bi − (Aj ∂j ) Bi
= [(B · ∇) A]i + [(∇ · B) A]i −
[(∇ · A) B]i − [(A · ∇) B]i
= [(B · ∇) A + (∇ · B) A − (∇ · A) B − (A · ∇) B]i
(Eq. 423)
(Eq. 494)
(ǫ is constant)
(product rule)
(Eq. 180)
(Eq. 554)
(distributivity)
(Eq. 555)
(grouping)
(Eqs. 413 & 419)
(Eq. 24)
Because i is a free index, the identity is proved for all components.
A·C A·D
:
B·C B·D
(A × B) · (C × D) = [A × B]i [C × D]i
= ǫijk Aj Bk ǫilm Cl Dm
= ǫijk ǫilm Aj Bk Cl Dm
= (δjl δkm − δjm δkl ) Aj Bk Cl Dm
= δjl δkm Aj Bk Cl Dm − δjm δkl Aj Bk Cl Dm
= (δjl Aj Cl ) (δkm Bk Dm ) − (δjm Aj Dm ) (δkl Bk Cl )
= (Al Cl ) (Bm Dm ) − (Am Dm ) (Bl Cl )
= (A · C) (B · D) − (A · D) (B · C)
• (A × B) · (C × D) =
=
A·C A·D
B·C B·D
• (A × B) × (C × D) = [D · (A × B)] C − [C · (A × B)] D:
[(A × B) × (C × D)]i = ǫijk [A × B]j [C × D]k
= ǫijk ǫjmn Am Bn ǫkpq Cp Dq
(Eq. 493)
(Eq. 494)
(commutativity)
(Eqs. 180 & 554)
(distributivity)
(commutativity)
(Eq. 555)
(Eq. 493)
(definition)
(Eq. 494)
(Eq. 494)
7.2 Tensors in Geometry
167
= ǫijk ǫkpq ǫjmn Am Bn Cp Dq
= ǫijk ǫpqk ǫjmn Am Bn Cp Dq
= (δip δjq − δiq δjp ) ǫjmn Am Bn Cp Dq
= (δip δjq ǫjmn − δiq δjp ǫjmn ) Am Bn Cp Dq
= (δip ǫqmn − δiq ǫpmn ) Am Bn Cp Dq
= δip ǫqmn Am Bn Cp Dq − δiq ǫpmn Am Bn Cp Dq
= ǫqmn Am Bn Ci Dq − ǫpmn Am Bn Cp Di
= ǫqmn Dq Am Bn Ci − ǫpmn Cp Am Bn Di
= (ǫqmn Dq Am Bn ) Ci − (ǫpmn Cp Am Bn ) Di
= [D · (A × B)] Ci − [C · (A × B)] Di
= [[D · (A × B)] C]i − [[C · (A × B)] D]i
= [[D · (A × B)] C − [C · (A × B)] D]i
(commutativity)
(Eq. 180)
(Eq. 554)
(distributivity)
(Eq. 555)
(distributivity)
(Eq. 555)
(commutativity)
(grouping)
(Eq. 495)
(index definition)
(Eq. 24)
Because i is a free index, the identity is proved for all components.
7.2
Tensors in Geometry
In this section, we give some examples of tensor calculus applications in geometry. These
examples are related mainly to the investigation of the properties of spaces in general and
space curvature in particular and hence they play important roles in non-Euclidean geometries and related applications in the physical sciences such as geometry-based gravitational
theories.
7.2.1
Riemann-Christoffel Curvature Tensor
This absolute rank-4 tensor, which is also called Riemann curvature tensor and RiemannChristoffel tensor, is a property of the space. It characterizes important properties of
spaces and hence it plays an important role in geometry in general and in non-Euclidean
geometries in particular with many applications in geometry-based physical theories. The
tensor is used, for instance, to test for the space flatness (see § 2.1). As indicated before,
the Riemann-Christoffel curvature tensor vanishes identically iff the space is globally flat
and hence the Riemann-Christoffel curvature tensor is zero in Euclidean spaces. The
Riemann-Christoffel curvature tensor depends only on the metric which, in general coordinate systems, is a function of position and hence the Riemann-Christoffel curvature
tensor follows this dependency on position. Yes, for affine coordinate systems the metric
tensor is constant and hence the Riemann-Christoffel curvature tensor vanishes identically.
There are two kinds of Riemann-Christoffel curvature tensor: first and second, where the
first kind is a type (0, 4) tensor while the second kind is a type (1, 3) tensor. The RiemannChristoffel curvature tensor of the first kind, which may also be called the covariant (or
totally covariant) Riemann-Christoffel curvature tensor, is defined by:
Rijkl = ∂k [jl, i] − ∂l [jk, i] + [il, r] Γrjk − [ik, r] Γrjl
(558)
7.2.1 Riemann-Christoffel Curvature Tensor
168
1
(∂k ∂j gli + ∂l ∂i gjk − ∂k ∂i gjl − ∂l ∂j gki ) + [il, r] Γrjk − [ik, r] Γrjl
2
1
(∂k ∂j gli + ∂l ∂i gjk − ∂k ∂i gjl − ∂l ∂j gki ) + g rs ([il, r] [jk, s] − [ik, r] [jl, s])
=
2
=
where the second step is based on Eq. 355 while the third step is based on Eq. 308. We
note that the first line of Eq. 558 can be cast in the following mnemonic determinantal
form:
Γrjk
Γrjl
∂k
∂l
Rijkl =
+
(559)
[jk, i] [jl, i]
[ik, r] [il, r]
Similarly, the Riemann-Christoffel curvature tensor of the second kind, which may also
be called the mixed Riemann-Christoffel curvature tensor, is given by:
Rijkl = ∂k Γijl − ∂l Γijk + Γrjl Γirk − Γrjk Γirl
(560)
We note again that Eq. 560 can be put into the following mnemonic determinantal form:
Rijkl =
∂k ∂l
Γijk Γijl
+
Γrjl Γrjk
Γirl Γirk
(561)
On lowering the contravariant index of the Riemann-Christoffel curvature tensor of the
second kind, the Riemann-Christoffel curvature tensor of the first kind is obtained, that
is:
Rijkl = gia Rajkl
(562)
Similarly, the Riemann-Christoffel curvature tensor of the second kind can be obtained by
raising the first index of the Riemann-Christoffel curvature tensor of the first kind, that
is:
Ri jkl = g ia Rajkl
(563)
One of the main applications of the Riemann-Christoffel curvature tensor in tensor calculus is demonstrated in its role in tensor differentiation. As seen in § 5.2, the covariant
differential operators in mixed derivatives are not commutative and hence for a covariant
vector A we have:
Aj;kl − Aj;lk = Rijkl Ai
(564)
From Eq. 564, it is obvious that the mixed second order covariant derivatives are equal
iff the Riemann-Christoffel curvature tensor of the second kind vanishes identically which
means that the space is flat and hence the covariant derivatives are ordinary partial derivatives. Similarly, for the mixed second order covariant derivative of a contravariant vector
A we have:
(565)
Aj;kl − Aj;lk = Rjilk Ai
which is similar to Eq. 564 for a covariant vector A. In brief, the covariant differential
operators become commutative when the metric makes the Riemann-Christoffel curvature
tensor of either kind vanish. So, since the Riemann-Christoffel curvature tensor is zero in
7.2.1 Riemann-Christoffel Curvature Tensor
169
Euclidean spaces, the mixed second order covariant derivatives, which become ordinary
partial derivatives, are equal when the C 2 continuity condition is satisfied.
The Riemann-Christoffel curvature tensor of the first kind satisfies the following symmetric and skew-symmetric relations in its four indices:
Rijkl =
Rklij
= −Rjikl
= −Rijlk
(block symmetry)
(anti-symmetry in the first two indices)
(anti-symmetry in the last two indices)
(566)
The skew-symmetric property of the covariant Riemann-Christoffel curvature tensor with
respect to the last two indices also applies to the mixed Riemann-Christoffel curvature
tensor, that is:
Rijkl = −Rijlk
(567)
As a consequence of the two anti-symmetric properties of the covariant Riemann-Christoffel
curvature tensor, the entries of the Riemann-Christoffel curvature tensor with identical values of the first two indices (e.g. Riijk ) or/and the last two indices (e.g. Rijkk ) are zero
(refer to § 3.1.5). Also, as a consequence of these anti-symmetric properties, all the entries
of the tensor with identical values of more than two indices (e.g. Riiji ) are zero.
We remark that all the above symmetric and anti-symmetric properties of the RiemannChristoffel curvature tensor of the first and second kinds can be proved by using the above
definitions of this tensor. For example, the first anti-symmetric property can be verified
as follows:
1
(∂k ∂i glj + ∂l ∂j gik − ∂k ∂j gil − ∂l ∂i gkj ) + g rs ([jl, r] [ik, s] − [jk, r] [il, s]) (568)
Rjikl =
2
1
rs
= −
(∂k ∂j gil + ∂l ∂i gkj − ∂k ∂i glj − ∂l ∂j gik ) + g ([jk, r] [il, s] − [jl, r] [ik, s])
2
1
sr
(∂k ∂j gli + ∂l ∂i gjk − ∂k ∂i gjl − ∂l ∂j gki ) + g ([il, s] [jk, r] − [ik, s] [jl, r])
= −
2
1
rs
(∂k ∂j gli + ∂l ∂i gjk − ∂k ∂i gjl − ∂l ∂j gki ) + g ([il, r] [jk, s] − [ik, r] [jl, s])
= −
2
= −Rijkl
where in the first line we use the third line of Eq. 558 with exchanging i and j, in the
second line we take a factor of −1, in the third line we use the symmetry of the metric
tensor in its two indices, in the fourth line we relabel two dummy indices, and in the last
line we use the third line of Eq. 558 again.
In an nD space, the Riemann-Christoffel curvature tensor has n4 components. As a
consequence of the symmetric and anti-symmetric properties of the covariant RiemannChristoffel curvature tensor, there are three types of distinct non-vanishing entries:
1. Entries with only two distinct indices (type Rijij ) which count:
N2 =
n (n − 1)
2
(569)
7.2.1 Riemann-Christoffel Curvature Tensor
170
2. Entries with only three distinct indices (type Rijik ) which count:
N3 =
n (n − 1) (n − 2)
2
(570)
3. Entries with four distinct indices (type Rijkl ) which count:
N4 =
n (n − 1) (n − 2) (n − 3)
12
(571)
The numerator in these three equations represents the k-permutations of n distinct objects
which is given by:
n!
P (n, k) =
(572)
(n − k)!
where k in these cases is equal to 2, 3, 4 respectively, while the denominator represents the
number of non-independent ways of generating these entries due to the anti-symmetric
and block symmetric properties.
By adding the numbers of the three types of non-zero distinct entries, as given by Eqs.
569-571, we can see that the Riemann-Christoffel curvature tensor in an nD space has a
total of:
n2 (n2 − 1)
NRI = N2 + N3 + N4 =
(573)
12
independent components which do not vanish identically. For example, in a 2D Riemannian space the Riemann-Christoffel curvature tensor has 24 = 16 components; however
there is only one independent component (with the principal suffix 1212) which is not
identically zero represented by the following four interdependent components:
R1212 = R2121 = −R1221 = −R2112
(574)
Similarly, in a 3D Riemannian space the Riemann-Christoffel curvature tensor has 34 = 81
components but only six of these are distinct non-zero entries which are the ones with the
following principal suffixes:
1212, 1313, 1213, 2123, 3132, 2323
(575)
where the permutations of the indices in each one of these suffixes are subject to the
symmetric and anti-symmetric properties of the four indices of the Riemann-Christoffel
curvature tensor, as in the case of a 2D space in the above example, and hence these
permutations do not produce independent entries. Following the pattern in the examples
of 2D and 3D spaces, in a 4D Riemannian space the Riemann-Christoffel curvature tensor
has 44 = 256 components but only 20 of these are independent non-zero entries, while in a
5D Riemannian space the Riemann-Christoffel curvature tensor has 54 = 625 components
but only 50 are independent non-zero entries.
On contracting the first covariant index with the contravariant index of the RiemannChristoffel curvature tensor of the second kind we obtain:
Riikl = ∂k Γiil − ∂l Γiik + Γril Γirk − Γrik Γirl
(j = i in Eq. 560)
(576)
7.2.2 Bianchi Identities
171
= ∂k Γiil − ∂l Γiik + Γril Γirk − Γirk Γril
∂k Γiil
(relabeling dummy i, r in last term)
∂l Γiik
−
=
√
√
= ∂k [∂l (ln g)] − ∂l [∂k (ln g)]
√
√
= ∂k ∂l (ln g) − ∂l ∂k (ln g)
√
√
= ∂k ∂l (ln g) − ∂k ∂l (ln g)
=0
(Eq. 324)
(C 2 continuity condition)
That is:
Riikl = 0
(577)
This is inline with the anti-symmetric property of the first two indices of the totally
covariant Riemann-Christoffel curvature tensor.
Finally, the Riemann-Christoffel curvature tensor also satisfies the following identity:
(578)
Rijkl;s + Riljk;s = Riksl;j + Rikjs;l
As indicated earlier, a necessary and sufficient condition for a space to have a coordinate
system such that all the components of the metric tensor are constants is that:
(579)
Rijkl = 0
This can be concluded from Eq. 558 plus Eqs. 307 and 308.
7.2.2
Bianchi Identities
The Riemann-Christoffel curvature tensor of the first and second kind satisfies a number
of identities called the Bianchi identities. The first Bianchi identity is given by:
Rijkl + Riljk + Riklj = 0
(first kind)
(580)
Rijkl + Riljk + Riklj = 0
(second kind)
(581)
These two forms of the first Bianchi identity can be obtained from each other by the index
raising and lowering operators. The first Bianchi identity, as stated above, is an instance
of the fact that by fixing the position of one of the four indices and permuting the other
three indices cyclically, the algebraic sum of these three permuting forms is zero, that is:
Rijkl + Riljk + Riklj
Rijkl + Rljik + Rkjli
Rijkl + Rlikj + Rjlki
Rijkl + Rkijl + Rjkil
=0
=0
=0
=0
(i fixed)
(j fixed)
(k fixed)
(l fixed)
(582)
(583)
(584)
(585)
All the above identities can be easily proved by using the definition of the RiemannChristoffel curvature tensor noting that the Christoffel symbols of both kinds are symmetric in their paired indices as given by Eqs. 310 and 311. For example, Eq. 580 can be
verified by substituting from the first line of Eq. 558 into Eq. 580, that is:
Rijkl + Riljk + Riklj = ∂k [jl, i] − ∂l [jk, i] + [il, r] Γrjk − [ik, r] Γrjl +
(586)
7.2.3 Ricci Curvature Tensor and Scalar
172
∂j [lk, i] − ∂k [lj, i] + [ik, r] Γrlj − [ij, r] Γrlk +
∂l [kj, i] − ∂j [kl, i] + [ij, r] Γrkl − [il, r] Γrkj
= 0
Another one of the Bianchi identities is:
Rijkl;m + Rijlm;k + Rijmk;l = 0
(first kind)
(587)
Rijkl;m + Rijlm;k + Rijmk;l = 0
(second kind)
(588)
Again, these two forms can be obtained from each other by the raising and lowering operators. We note that the pattern of the second Bianchi identity in its both kinds is simple,
that is the first two indices are fixed while the last three indices are cyclically permuted
in the three terms. It is noteworthy that the Bianchi identities are valid regardless of the
space metric.
7.2.3
Ricci Curvature Tensor and Scalar
The Ricci curvature tensor, which is a rank-2 absolute symmetric tensor, is a byproduct of
the Riemann-Christoffel curvature tensor and hence it plays a similar role in characterizing
the space and describing its curvature. There are two kinds of Ricci curvature tensor: first
and second, where the first kind is a type (0, 2) tensor while the second kind is a type (1, 1)
tensor. The first kind of this tensor is obtained by contracting the contravariant index
with the last covariant index of the Riemann-Christoffel curvature tensor of the second
kind, that is:
(589)
Rij = Raija = ∂j Γaia − ∂a Γaij + Γabj Γbia − Γaba Γbij
where Eq. 560 is used in this formulation. The Ricci curvature tensor, as given by the
last equation, can be written in the following mnemonic determinantal form:
Γabj Γaba
∂j ∂a
(590)
+
Γaij Γaia
Γbij Γbia
√
Because of Eq. 324 (i.e. Γjij = ∂i ln g ), the Ricci tensor can also be written in the
following forms as well as several other forms:
√
√
(591)
Rij = ∂j ∂i (ln g) − ∂a Γaij + Γabj Γbia − Γbij ∂b (ln g)
1
√
√ a
= ∂j ∂i (ln g) + Γabj Γbia − √ ∂a
gΓij
g
Rij =
where g is the determinant of the covariant metric tensor. As stated above, the Ricci
tensor of the first kind is symmetric, that is:
Rij = Rji
(592)
This can be easily verified by exchanging the i and j indices in the last line of Eq. 591
taking account of the C 2 continuity condition and the fact that the Christoffel symbols
are symmetric in their paired indices (Eq. 311).
7.3 Tensors in Science
173
On raising the first index of the Ricci tensor of the first kind, the Ricci tensor of the
second kind is obtained, that is:
Rij = g ik Rkj
(593)
This process can be reversed and hence the Ricci tensor of the first kind can be obtained
from the Ricci tensor of the second kind using the index lowering operator. In an nD
space, the Ricci tensor has n2 entries. However, because of its symmetry the number of
its distinct entries is reduced to:
n (n + 1)
(594)
NRD =
2
The Ricci scalar R, which is also called the curvature scalar and the curvature invariant,
is the result of contracting the indices of the Ricci tensor of the second kind, that is:
R = Rii
(595)
Since the Ricci scalar is obtained by raising a covariant index of the Ricci tensor of the
first kind using the raising operator followed by contracting the two indices, it may be
written by some in the following form:
1
√ a
√
a b
ij
ij
(596)
gΓij
R = g Rij = g ∂j ∂i (ln g) + Γbj Γia − √ ∂a
g
where the expression inside the square brackets is obtained from the last line of Eq. 591.
Similar expressions can be obtained from the other lines of that equation.
7.3
Tensors in Science
In this section, we give a few examples of tensor calculus applications in science. These
examples come mainly from the disciplines of continuum mechanics and fluid dynamics.
For simplicity, clarity and widespread use we employ a Cartesian approach in the following
formulations.
7.3.1
Infinitesimal Strain Tensor
This is a rank-2 tensor which describes the state of strain in a continuum medium and
hence it is used in continuum and fluid mechanics. The infinitesimal strain tensor γ is
defined by:
∇d + (∇d)T
γ=
(597)
2
where d is the displacement vector and the superscript T represents matrix transposition.
We note that some authors do not include the factor 21 in the definition of γ. The displacement vector d represents the change in distance and direction which an infinitesimal
element of the medium experiences as a consequence of the applied stress. In Cartesian
coordinates, the last equation is given in tensor notation by the following form:
γij =
∂i dj + ∂j di
2
(598)
7.3.2 Stress Tensor
7.3.2
174
Stress Tensor
The stress tensor, which is also called Cauchy stress tensor, is a rank-2 tensor used for
transforming a normal vector to a surface to a traction vector acting on that surface, that
is:
T=σ·n
(599)
where T is the traction vector, σ is the stress tensor and n is the normal vector. This is
usually expressed in tensor notation using Cartesian coordinates by the following form:
(600)
Ti = σij nj
We should remark that the stress tensor is symmetric in many applications (e.g. in the
flow of Newtonian fluids) but not all, as it can be asymmetric in some cases. We also
remark that we chose to define the stress tensor within the context of Cauchy stress law
which is more relevant to the continuum mechanics, although it can be defined differently
in other disciplines and in a more general form.
The diagonal components of the stress tensor represent normal stresses while the offdiagonal components represent shear stresses. Assuming that the stress tensor is symindependent components, instead of n2 , where n is the space
metric, it possesses n(n+1)
2
dimension. Hence in a 3D space (which is the natural space for this tensor in the common
physical applications) it has six independent components. In fluid dynamics, the stress
tensor (which may also be labeled as the total stress tensor) is decomposed into two main
parts: a viscous part and a pressure part. The viscous part may then be split into a
normal stress and a shear stress while the pressure part may be split into a hydrostatic
pressure and an extra pressure of hydrodynamic nature.
7.3.3
Displacement Gradient Tensors
These are rank-2 tensors denoted by E and ∆. They are defined in Cartesian coordinates
using tensor notation as:
′
∂xi
∆ij =
∂xj
∂xi
Eij =
′
∂xj
(601)
′
where the indexed x and x represent the Cartesian coordinates of an observed continuum
particle at the present and past times respectively. These tensors may also be called
the deformation gradient tensors. The first displacement gradient tensor E quantifies the
displacement of a particle at the present time relative to its position at the past time,
while the second displacement gradient tensor ∆ quantifies its displacement at the past
time relative to its position at the present time. From their definitions, it is obvious that
E and ∆ are inverses of each other and hence:
Eik ∆kj = δij
(602)
7.3.4 Finger Strain Tensor
7.3.4
175
Finger Strain Tensor
This rank-2 tensor, which may also be called the left Cauchy-Green deformation tensor, is
used in the fields of fluid and continuum mechanics to describe the strain in a continuum
object, e.g. fluid, in a series of time frames. It is defined as:
B = E · ET
(603)
where E is the first displacement gradient tensor as defined in § 7.3.3 and the superscript
T represents matrix transposition. The last equation can be expressed in tensor notation,
using Cartesian coordinates, as follows:
Bij =
∂xi ∂xj
′
′
∂xk ∂xk
(604)
′
where the indexed x and x represent the Cartesian coordinates of an element of the
continuum at the present and past times respectively.
7.3.5
Cauchy Strain Tensor
This tensor, which may also be called the right Cauchy-Green deformation tensor, is
the inverse of the Finger strain tensor (see § 7.3.4) and hence it is denoted by B−1 .
Consequently, it is defined as:
B−1 = ∆T · ∆
(605)
where ∆ is the second displacement gradient tensor as defined in § 7.3.3. The last equation
can be expressed in tensor notation, using Cartesian coordinates, as follows:
′
Bij−1
′
∂xk ∂xk
=
∂xi ∂xj
(606)
The Finger and Cauchy strain tensors may be called “finite strain tensors” as opposite
to infinitesimal strain tensor (see § 7.3.1). They are symmetric positive definite tensors;
moreover they become the unity tensor when the change in the state of the object from
the past to the present times consists of rotation and translation with no deformation.
7.3.6
Velocity Gradient Tensor
This is a rank-2 tensor which is often used in fluid dynamics and rheology. As its name
suggests, it is the gradient of the velocity vector v and hence it is given in Cartesian
coordinates by:
[∇v]ij = ∂i vj
(607)
The velocity gradient tensor in other coordinate systems can be obtained from the expression of the gradient of vectors in these systems, as given, for instance, in § 6.4 and § 6.5
for cylindrical and spherical coordinates. The term “velocity gradient tensor” may also be
used for the transpose of this tensor, i.e. (∇v)T . The velocity gradient tensor is usually
7.3.7 Rate of Strain Tensor
176
decomposed into a symmetric part which is the rate of strain tensor S (see § 7.3.7), and
an anti-symmetric part which is the vorticity tensor S̄ (see § 7.3.8), that is:
∇v = S + S̄
(608)
As seen earlier (refer to § 3.1.5), any rank-2 tensor can be decomposed into a symmetric
part and an anti-symmetric part.
7.3.7
Rate of Strain Tensor
This rank-2 tensor, which is also called the rate of deformation tensor, is the symmetric
part of the velocity gradient tensor and hence it is given by:
∇v + (∇v)T
2
which, in tensor notation with Cartesian coordinates, becomes (see Eq. 99):
S=
(609)
∂ i vj + ∂ j vi
(610)
2
We note that some authors do not include the factor 21 in the definition of S and S̄ and
hence this factor is moved to the definition of ∇v (Eq. 608). Also, the tensors S and S̄
are commonly denoted by γ̇ and ω respectively.
The rate of strain tensor is a quantitative measure of the local rate at which neighboring
material elements of a deforming continuum move with respect to each other. As a rank-2
symmetric tensor, it has n(n+1)
independent components which are six in a 3D space. The
2
rate of strain tensor is related to the infinitesimal strain tensor (refer to § 7.3.1) by:
Sij =
∂γ
(611)
∂t
where t is time. Hence, the rate of strain tensor is normally denoted by γ̇ where the dot
represents the temporal rate of change, as indicated above.
S=
7.3.8
Vorticity Tensor
This rank-2 tensor is the anti-symmetric part of the velocity gradient tensor and hence it
is given by:
∇v − (∇v)T
(612)
S̄ =
2
which, in tensor notation with Cartesian coordinates, becomes (see Eq. 100):
∂ i vj − ∂ j vi
(613)
2
The vorticity tensor quantifies the local rate of rotation of a deforming continuum medium.
As a rank-2 anti-symmetric tensor, it has n(n−1)
independent non-zero components which
2
are three in a 3D space. These three components, added to the aforementioned six components of the rate of strain tensor (see § 7.3.7), give nine independent components which
is the total number of components of their parent tensor ∇v.
S̄ij =
7.4 Exercises and Revision
7.4
177
Exercises and Revision
7.1 Summarize the reasons for the popularity of tensor calculus techniques in mathematical, scientific and engineering applications.
7.2 State, in tensor language, the definition of the following mathematical concepts assuming Cartesian coordinates of a 3D space: trace of matrix, determinant of matrix,
inverse of matrix, multiplication of two compatible square matrices, dot product of
two vectors, cross product of two vectors, scalar triple product of three vectors and
vector triple product of three vectors.
7.3 From the tensor definition of A×(B × C), obtain the tensor definition of (A × B)×C.
7.4 We have the following tensors in orthonormal Cartesian coordinates of a 3D space:
A = (22, 3π, 6.3)
B = (3e, 1.8, 4.9)
C = (47, 5e, 3.5)
3 e2
π 3
E=
D=
π3 7
4 e
7.5
7.6
7.7
7.8
Use the tensor expressions for the relevant mathematical concepts with systematic
substitution of the indices values to find the following: tr (D), det (E), D−1 , E · D,
A · C, C × B, C · (A × B) and B × (C × A).
State the matrix and tensor definitions of the main three independent scalar invariants
(I, II and III) of rank-2 tensors.
Express the main three independent scalar invariants (I, II and III) of rank-2 tensors
in terms of the three subsidiary scalar invariants (I1 , I2 and I3 ).
Referring to question 7.4, find the three scalar invariants (I, II and III) of D and
the three scalar invariants (I1 , I2 and I3 ) of E using the tensor definitions of these
invariants with systematic index substitution.
State the following vector identities in tensor notation:
∇×r
∇ · (f A)
A × (∇ × B)
∇ × (A × B)
=
=
=
=
0
f ∇ · A + A · ∇f
(∇B) · A − A · ∇B
(B · ∇) A + (∇ · B) A − (∇ · A) B − (A · ∇) B
7.9 State the divergence and Stokes theorems for a vector field in Cartesian coordinates
using vector and tensor notations. Also, define all the symbols involved.
7.10 Prove the following vector identities using tensor notation and techniques with full
justification of each step:
∇·r
∇ · (∇ × A)
A · (B × C)
∇ × (∇ × A)
=
=
=
=
n
0
C · (A × B)
∇ (∇ · A) − ∇2 A
7.4 Exercises and Revision
178
7.11 What is the type, in the form of (m, n, w), of the Riemann-Christoffel curvature tensor
of the first and second kinds?
7.12 What are the other names used to label the Riemann-Christoffel curvature tensor of
the first and second kinds?
7.13 What is the importance of the Riemann-Christoffel curvature tensor with regard to
characterizing the space as flat or curved?
7.14 State the mathematical definition of the Riemann-Christoffel curvature tensor of either
kinds in determinantal form.
7.15 How can we obtain the Riemann-Christoffel curvature tensor of the first kind from
the second kind and vice versa?
7.16 Using the definition of the second order mixed covariant derivative of a vector field
(see § 5.2) and the definition of the mixed Riemann-Christoffel curvature tensor, verify
the following equation: Aj;kl −Aj;lk = Rijkl Ai . Repeat the question with the equation:
Aj;kl − Aj;lk = Rjilk Ai .
7.17 Based on the equations in question 7.16, what is the necessary and sufficient condition
for the covariant differential operators to become commutative?
7.18 State, mathematically, the anti-symmetric and block symmetric properties of the
Riemann-Christoffel curvature tensor of the first kind in its four indices.
7.19 Based on the two anti-symmetric properties of the covariant Riemann-Christoffel curvature tensor, list all the forms of the components of the tensor that are identically
zero (e.g. Riijk ).
7.20 Verify the block symmetric property and the two anti-symmetric properties of the
covariant Riemann-Christoffel curvature tensor using its definition.
7.21 Repeat question 7.20 for the anti-symmetric property of the mixed Riemann-Christoffel
curvature tensor in its last two indices.
7.22 Based on the block symmetric and anti-symmetric properties of the covariant RiemannChristoffel curvature tensor, find (with full justification) the number of distinct nonvanishing entries of the three main types of this tensor (see Eqs. 569-571). Hence,
find the total number of the independent non-zero components of this tensor.
7.23 Use the formulae found in question 7.22 and other formulae given in the text to find
the number of all components, the number of non-zero components, the number of
zero components and the number of independent non-zero components of the covariant
Riemann-Christoffel curvature tensor in 2D, 3D and 4D spaces.
7.24 Prove the following identity with full justification of each step of your proof: Raakl = 0.
7.25 Make a list of all the main properties of the Riemann-Christoffel curvature tensor (i.e.
rank, type, symmetry, etc.).
7.26 Prove the following identity using the Bianchi identities: Rijkl;s + Riljk;s = Riksl;j +
Rikjs;l .
7.27 Write the first Bianchi identity in its first and second kinds.
7.28 Verify the following form of the first Bianchi identity using the mathematical definition
of the Riemann-Christoffel curvature tensor: Rijkl + Rkijl + Rjkil = 0.
7.29 What is the pattern of the indices in the second Bianchi identity?
7.30 Write the determinantal form of the Ricci curvature tensor of the first kind.
7.4 Exercises and Revision
179
7.31 Starting from the determinantal form of the Ricci curvature tensor of the first kind,
obtain the following form of the Ricci curvature tensor with justification
of each step
√ a
√
in your derivation: Rij = ∂j ∂i ln g + Γabj Γbia − √1g ∂a
gΓij .
7.32 Verify the symmetry of the Ricci tensor of the first kind in its two indices.
7.33 What is the number of distinct entries of the Ricci curvature tensor of the first kind?
7.34 How can we obtain the Ricci curvature scalar from the covariant Riemann-Christoffel
curvature tensor? Write an orderly list of all the required steps to do this conversion.
7.35 Make a list of all the main properties of the Ricci curvature tensor (rank, type, symmetry, etc.) and the Ricci curvature scalar.
7.36 Outline the importance of the Ricci curvature tensor and the Ricci curvature scalar
in characterizing the space.
7.37 Write, in tensor notation, the mathematical expressions of the following tensors in
Cartesian coordinates defining all the symbols involved: infinitesimal strain tensor,
stress tensor, first and second displacement gradient tensors, Finger strain tensor,
Cauchy strain tensor, velocity gradient tensor, rate of strain tensor and vorticity
tensor.
7.38 Which of the tensors in question 7.37 are symmetric, anti-symmetric or neither?
7.39 Which of the tensors in question 7.37 are inverses of each other?
7.40 Which of the tensors in question 7.37 are derived from other tensors in that list?
7.41 What is the relation between the first and second displacement gradient tensors?
7.42 What is the relation between the velocity gradient tensor and the rate of strain tensor?
7.43 What is the relation between the velocity gradient tensor and the vorticity tensor?
7.44 What is the relation between the rate of strain tensor and the infinitesimal strain
tensor?
7.45 What are the other names given to the following tensors: stress tensor, deformation
gradient tensors, left Cauchy-Green deformation tensor, Cauchy strain tensor and rate
of deformation tensor?
7.46 What is the physical significance of the following tensors: infinitesimal strain tensor,
stress tensor, first and second displacement gradient tensors, Finger strain tensor,
velocity gradient tensor, rate of strain tensor and vorticity tensor?
References
G.B. Arfken; H.J. Weber; F.E. Harris. Mathematical Methods for Physicists A Comprehensive Guide. Elsevier Academic Press, seventh edition, 2013.
R.B. Bird; R.C. Armstrong; O. Hassager. Dynamics of Polymeric Liquids, volume 1.
John Wiley & Sons, second edition, 1987.
R.B. Bird; W.E. Stewart; E.N. Lightfoot. Transport Phenomena. John Wiley & Sons,
second edition, 2002.
M.L. Boas. Mathematical Methods in the Physical Sciences. John Wiley & Sons Inc.,
third edition, 2006.
J. Bonet; R.D. Wood. Nonlinear Continuum Mechanics for Finite Element Analysis.
Cambridge University Press, first edition, 1997.
C.F. Chan Man Fong; D. De Kee; P.N. Kaloni. Advanced Mathematics for Engineering
and Science. World Scientific Publishing Co. Pte. Ltd., first edition, 2003.
T.L. Chow. Mathematical Methods for Physicists: A concise introduction. Cambridge
University Press, first edition, 2003.
P. Grinfeld. Introduction to Tensor Analysis and the Calculus of Moving Surfaces.
Springer, first edition, 2013.
J.H. Heinbockel. Introduction to Tensor Calculus and Continuum Mechanics. 1996.
D.C. Kay. Schaum’s Outline of Theory and Problems of Tensor Calculus. McGraw-Hill,
first edition, 1988.
K.F. Riley; M.P. Hobson; S.J. Bence. Mathematical Methods for Physics and Engineering.
Cambridge University Press, third edition, 2006.
T. Sochi. Tensor Calculus Made Simple. CreateSpace, first edition, 2016.
T. Sochi. Introduction to Differential Geometry of Space Curves and Surfaces. CreateSpace, first edition, 2017.
I.S. Sokolnikoff. Tensor Analysis Theory and Applications. John Wiley & Sons, Inc., first
edition, 1951.
B. Spain. Tensor Calculus: A Concise Course. Dover Publications, third edition, 2003.
J.L. Synge; A. Schild. Tensor Calculus. Dover Publications, 1978.
D. Zwillinger, editor. CRC Standard Mathematical Tables and Formulae. CRC Press,
32nd edition, 2012.
180
Index
Absolute
covariant derivative, 130
derivative, 6, 11, 94, 109, 129–133, 136
differentiation, 94, 112, 129–132, 136
permutation tensor, 8, 81, 90, 99, 100, 107, 110
tensor, 57, 58, 72, 76, 81, 88, 89, 91, 107, 108,
167, 172
value, 11, 81
Active transformation, 33
Addition of tensors, 14, 62, 63, 73
Admissible transformation, 26, 31, 32, 34, 41, 47–49,
62, 112, 141
Affine
coordinate system, 25, 30, 118, 167
tensor, 62, 112, 120, 132, 133
transformation, 30
Algebraic
addition, 34, 73
subtraction, 73
Anisotropic tensor, 58, 72
Antisymmetric, 50, 59–62, 72, 73, 77, 78, 81, 84,
107, 169–171, 176, 178, 179
symmetric tensor, 59–62, 73, 176
symmetry, 59, 61, 73, 107, 169
Area, 6, 39, 76, 103–105, 110, 160
Associate
metric tensor, 91
tensor, 94, 109
Associative, 63–65, 73
Associativity, 32, 63, 65
Asymmetric tensor, 59, 174
Axial tensor, 55
Basis
tensor, 54, 71–73, 141
vector, 6, 7, 9, 15, 17, 24–28, 32, 39–46, 48,
49, 53–55, 69–72, 74, 76, 77, 91–94, 97, 99,
100, 109, 112–114, 120, 121, 123–125, 128–
130, 132, 134–136, 140, 141, 146, 148, 150,
152
Bianchi identity, 171, 172, 178
Block
symmetric, 170, 178
symmetry, 169
Bound index, 14, 20
Cartesian, 1, 6, 7, 10, 15, 16, 22–30, 33–36, 39, 42,
47, 48, 51, 53–56, 66, 68, 69, 71, 72, 74,
83, 91–95, 98, 99, 108–111, 113, 118, 119,
121, 134, 137, 138, 140, 146, 152, 155, 161,
173–177, 179
Cauchy
-Green deformation tensor, 175, 179
strain tensor, 6, 175, 179
stress law, 174
stress tensor, 174
Chain rule, 130, 131
Christoffel symbol, 6, 7, 94, 112–123, 128, 129, 132–
135, 144, 153, 171, 172
Circle, 25, 26, 35, 36
Closure, 32
Comma notation, 6, 10, 11, 19, 117
Commutative, 11, 32, 48, 63–65, 67, 73, 74, 85, 92,
124, 131, 135, 168, 178
Commutativity, 63, 84, 93, 126, 163–167
Complex, 9, 11
Composition of transformations, 32, 47
Conjugate
metric tensor, 91
tensor, 94, 109
Conserved, 77, 78, 81, 107
Continuity condition, 11, 20, 31, 47, 124, 162, 169,
171, 172
Continuous, 11, 19, 22, 30, 31, 34, 127
Continuum mechanics, 69, 155, 173–175
Contracted epsilon identity, 87
Contraction, 14, 57, 62, 66–69, 73, 74, 82, 96, 124,
126, 132, 135, 137, 139, 144, 152
Contravariant, 7
basis vector, 6, 24–26, 39–41, 43–45, 48, 49, 53–
55, 69–71, 74, 91, 92, 109, 113, 132, 140
component, 43, 48, 49, 54, 69, 71, 97, 100, 102,
121, 144
differentiation, 121, 134
index, 51, 52, 58, 61, 93, 120, 122, 142–144,
168, 170, 172
Kronecker delta, 132
metric tensor, 7, 39, 43, 52, 70, 92, 93, 96, 109,
110, 114, 124, 125, 132, 142
partial derivative, 11
permutation tensor, 8, 84, 100, 107, 132
tensor, 32, 50–55, 65, 99, 121, 143, 153
tensor derivative, 11
181
Coordinate
curve, 24–27, 35–40, 43–48, 53, 55, 103–106,
123
surface, 24–26, 35–40, 43–46, 48, 53, 55, 105,
110, 111
system, 7, 10, 12, 13, 15, 16, 20, 22–36, 39–
41, 44–51, 53, 55, 57, 58, 62, 70–72, 74–77,
83, 91–95, 97, 99, 102, 103, 105, 108–113,
118–121, 123–125, 128, 129, 132–138, 140,
144–148, 150, 152, 158, 161, 167, 171, 175
Covariant
basis vector, 6, 24–27, 39–41, 43–45, 48, 49,
53–55, 69–71, 74, 91, 92, 99, 109, 113, 114,
132
component, 43, 48, 49, 54, 69, 71, 97, 100, 101,
121, 141, 143
derivative, 6, 11, 120–131, 133–136, 141–143,
145, 168, 178
differentiation, 10, 112, 120–132, 134–136
index, 51–53, 58, 93, 122, 137, 152, 170, 172,
173
Kronecker delta, 132
metric tensor, 7, 39, 42, 43, 52, 53, 70, 81, 92–
94, 96, 99, 103, 105, 108–110, 114, 124,
125, 132, 142, 172
permutation tensor, 8, 99, 132
Riemann-Christoffel curvature tensor, 167, 169,
171, 178
tensor, 32, 50–55, 64, 65, 67, 99, 110, 121, 145
Cross product, 43, 56, 57, 84, 99, 100, 103, 104, 110,
137, 139, 143, 152, 157, 163, 164, 177
Curl, 6, 56, 72, 139, 143, 144, 147, 149, 151–154
Curvature
invariant, 173
scalar, 173
Curve, 6, 7, 11, 12, 23–25, 35, 36, 53, 76, 98, 99,
103, 110, 129–132, 136
Curved space, 22–24, 46, 94, 109, 110, 178
Curvilinear coordinate system, 24–26, 30, 34, 36,
40, 47, 48, 71, 120, 121, 123, 128, 133, 134,
137
Cylindrical coordinate system, 6, 8, 24–26, 28, 35,
37, 39, 47, 48, 71, 95, 110, 111, 113, 117–
119, 134, 137, 148, 150, 152, 153, 175
calculus, 11
geometry, 12, 17, 24, 119, 155
operator, 6, 10, 11, 51, 52, 121, 124, 126, 135,
137, 138, 140, 146, 152, 168, 178
Dimension of space, 6, 12, 15–18, 20, 24, 29, 30, 39,
44, 46, 58, 64, 65, 72, 76, 77, 82, 83, 107,
119, 156, 158, 174
Direct
multiplication, 64–66
notation, 12
product, 56–58, 64, 65, 73
transformation, 29
Displacement
gradient tensor, 6, 7, 174, 175, 179
vector, 6, 50, 102–106, 173
Distributive, 18, 64, 65, 67, 73
Distributivity, 163–165, 167
Divergence, 6, 137, 139, 142–145, 147, 149, 151–154
theorem, 160, 177
Dot product, 27, 42, 43, 53, 66–68, 74, 92, 93, 97,
98, 110, 114, 129, 130, 139, 156, 177
Double inner product, 68, 83, 84
Dummy index, 9, 14–18, 20, 52, 169
Dyad, 6, 55, 56, 65, 68, 72, 83, 122, 148–151
Dyadic
multiplication, 64
product, 54
Ellipsoid, 23, 94
epsilon-delta identity, 87, 90, 108
Euclidean, 22–24, 27, 46, 94, 134, 167, 169
Exterior multiplication, 64
Extra pressure, 174
Finger strain tensor, 6, 175, 179
Flat
metric, 94
space, 23, 27, 31, 46, 47, 94, 96, 109, 110, 127,
167, 168, 178
Fluid
dynamics, 138, 173–175
mechanics, 69, 155, 173, 175
Form-invariance, 77
Form-invariant, 12, 13
Free index, 9, 14–18, 20, 51, 52, 59, 63, 66, 123,
161–167
Determinant, 6, 7, 11, 29, 30, 42, 43, 53, 72, 81, 85–
89, 91, 92, 94, 99, 102, 105, 108–110, 115,
Gauss theorem, 155
126, 142, 156, 158, 172, 177
General
Diagonal matrix, 44, 95
coordinate system, 1, 7, 10, 11, 20, 22, 24, 27,
Differentiable, 31, 121–124, 127, 129–131, 134–136,
39, 41–43, 53, 57, 66, 70, 72–74, 91–93, 97,
138–144, 147, 149–152, 158, 160, 161
99, 100, 102, 103, 105, 106, 108–112, 119,
Differential
182
120, 122, 133, 136, 137, 140, 141, 144–146,
tensor, 58, 72, 76–78, 81
148, 150, 152, 153, 161, 167
Jacobian, 7, 29–33, 39, 42, 47, 48, 57, 94, 105
tensor, 62, 112, 120, 126, 133, 136, 143
matrix, 7, 29–31, 33, 42, 47
Generalized Kronecker delta, 7, 76, 82, 88–90, 108,
126
Kronecker delta, 7, 15, 41, 53, 58, 71, 76, 77, 82,
Gibbs notation, 12
85–90, 93, 96, 107–109, 126, 127, 132, 135,
Gradient, 6, 32, 39, 43, 45, 50, 129, 136–142, 144,
136
145, 147, 149, 150, 152–154, 175
vector, 32, 43
Laplacian, 6, 10, 137, 138, 140, 144–146, 148–154,
164
Handedness, 32, 55, 81, 105
Length, 6, 7, 15, 22, 23, 25, 27, 29, 39, 46, 53, 76,
Homogeneous coordinate system, 23, 27, 29, 47, 110
83, 91, 92, 94, 96, 102, 103, 108, 110, 157
Hydrostatic pressure, 174
Levi-Civita
identity, 87
Identity, 32
tensor, 58, 77
tensor, 22, 43, 76, 83, 91, 94
Linear
transformation, 34, 48
algebra, 42, 44, 67, 74, 95, 155, 156
Imaginary, 9, 11, 27, 29, 92, 96
combination, 113
Improper
coordinate system, 25
rotation, 58
operation, 124
transformation, 32–34, 48, 55, 77, 78, 81, 107
transformation, 26, 27, 29, 34, 62, 112, 121
Index
Lorentz
transformations, 23, 29, 95, 110
-free notation, 12, 13
Lowering
operator, 52, 71, 93, 94, 97, 112–114, 126,
notation, 9
168,
171–173
Indicial
notation, 12, 13, 17, 19, 126, 135
structure, 17, 18, 20, 51, 52, 58, 63, 64
Infinitesimal, 6, 22, 23, 34, 46, 91, 103–106, 108,
111, 173
strain tensor, 7, 173, 175, 176, 179
Inner
multiplication, 73, 74, 94, 113
product, 6, 57, 58, 64–69, 74, 83, 84, 94, 97, 98,
113, 124, 126, 127, 129, 131, 156
Integral, 105, 110
theorem, 155, 160, 161
Intrinsic derivative, 11, 130, 136
Invariance, 12, 16, 19, 71–73, 91, 112, 120
Invariant, 12, 13, 31, 32, 41, 47, 55, 58, 61, 62, 73,
76, 77, 98, 103, 121, 139, 141, 153, 155,
157, 158, 177
Inverse, 32
Jacobian, 7, 30, 31
of matrix, 44, 95, 156, 177
of metric, 44, 91, 92, 95, 109
of tensor, 174, 175, 179
transformation, 29–31, 47
Invertible
matrix, 43, 91, 95, 158
transformation, 30, 31
Irrotational, 160
Isotropic, 23
Manifold, 22–24, 31, 39, 42, 46, 94
Matrix
algebra, 65, 66, 74, 155, 156
notation, 156
Metric
space, 22
tensor, 7, 15, 22, 23, 26, 27, 32, 39, 40, 42–
49, 52, 53, 69–71, 76, 81, 90–99, 103, 105,
107–110, 112–114, 118, 119, 123–126, 129,
131–136, 142, 147, 148, 150, 167–169, 171,
172
Minkowski
metric, 95, 110
space, 23, 29
Mixed
derivative, 10, 11, 127, 128, 135, 168, 169, 178
Kronecker delta, 7, 77, 82, 93, 132
metric tensor, 7, 22, 41, 43, 91–93, 109, 124,
125, 132
Riemann-Christoffel curvature tensor, 168, 169,
178
tensor, 7, 14, 18–22, 32, 41, 43, 51, 52, 54, 63,
65, 66, 69, 71, 77, 81, 82, 91–94, 102, 109,
121, 122, 124, 125, 131, 132, 141, 143, 152,
168, 169, 178
Multiplication
by scalar, 62, 64, 73
183
of matrices, 67, 73, 92, 156, 177
product, 6, 64–68, 94, 124, 126, 131, 138
of matrix by vector, 67, 156
of tensors, 14, 58, 62, 64–67, 73, 74, 89, 94, 113, Parallelepiped, 41, 105
Partial
124, 131
derivative, 6, 10, 11, 30, 31, 53, 96, 112–114,
Mutually
117, 118, 120–124, 126–130, 132, 133, 135,
exclusive, 50
168, 169
independent, 24, 40, 44
differential
operator, 11, 51, 52, 121, 138
orthogonal, 15, 25, 26, 36, 39, 43, 44, 49, 53,
differentiation,
10, 83, 96, 120, 121, 124, 134,
83, 92
135
perpendicular, 24–27, 55, 94
Passive transformation, 33
nabla operator, 6, 137–140, 143, 146, 148, 150, 152, Perimeter, 161
Permissible transformation, 31, 62, 94, 98
153
Permutation
Negative orthogonal transformation, 34, 48
of tensor, 68, 69, 74
Nontensor, 8, 57, 76–78, 80–90, 99, 100, 102, 107,
Euclidean, 167
110, 132, 139
scalar tensor, 24, 32, 64, 65, 112, 120, 121, 128,
Perpendicular, 6, 24–27, 39, 53, 55, 94
133, 134, 136, 146
Physical
singular, 91, 92
basis vector, 71, 74
Nonlinear transformation, 26, 34
component, 69–71, 75, 137, 147, 148, 150, 153
Normal
dimension, 25, 46, 69
stress, 174
representation, 6, 39, 69, 71, 74
vector, 7, 174
Plane, 23, 25, 26, 35, 36, 46, 55, 94
Normalized vector, 6, 102
Polar
Oblique coordinate system, 25, 26
angle, 36
Order
coordinate system, 8, 36, 148
of derivative, 10, 11, 127, 128, 135, 168, 169,
tensor, 55
178
Polyad, 55
of indices, 10, 16–21, 51, 52, 55, 64, 66, 83, 93, Position vector, 6, 7, 39, 53, 102, 146, 158, 161
94, 109, 119, 123, 134, 141, 162
Positive orthogonal transformation, 33, 34, 48
of multiplicands, 67, 73, 124, 131, 157, 162–164 Principle of invariance, 12, 19, 112
of operators, 11, 18, 138, 162
Product rule, 96, 114, 121–125, 130, 132, 141, 142,
of tensor, 14, 16, 18, 20
145, 162, 163, 165, 166
of transformations, 32
Proper
Orthogonal, 15, 25–27, 36, 39, 43, 44, 49, 53, 83, 92
rotation, 58
coordinate system, 7, 24–26, 36, 39, 41, 43, 47,
tensor, 55
70, 71, 74, 94, 95, 103–105, 108–111, 115–
transformation, 32–34, 48, 77, 78, 81, 107
117, 119, 133, 134, 137, 146–148, 150, 153 Pseudo
transformation, 27, 33, 34, 48, 55, 78, 81
tensor, 55–58, 63, 72, 78, 81
Orthonormal
vector, 55, 57, 72, 157
basis set, 15, 43, 49, 53, 71, 83, 148, 150
Cartesian, 15, 16, 22, 23, 27, 28, 30, 39, 48, 51, Quotient rule of
differentiation, 69
55, 66, 68, 71, 74, 83, 91, 93–95, 98, 99,
tensor, 62, 69, 74
108–111, 118, 119, 134, 177
coordinate system, 15, 16, 20, 22, 23, 27, 28,
Radius, 23
30, 39, 48, 51, 55, 66, 68, 71, 74, 83, 91,
Raising operator, 52, 71, 93, 94, 97, 112, 121, 126,
93–95, 98, 99, 108–111, 118, 119, 134, 177
141, 142, 153, 168, 171–173
vectors, 6, 15, 20, 49, 72, 83, 84, 99, 108
Range of index, 12–16, 18, 65, 79, 80, 160
Orthonormalized vectors, 6, 53, 72
Rank
Outer
-0 tensor, 9, 12, 13, 16, 19, 58, 62, 65, 67, 153,
multiplication, 64, 67, 73
157
184
-1 tensor, 9, 12–14, 19, 50, 56, 58, 64, 67, 70,
78, 140, 152, 157
-2 tensor, 9, 10, 12–14, 16, 19, 22, 54, 56, 58–
60, 64–68, 70–74, 76, 78, 79, 85, 86, 91,
97, 102, 107, 110, 121, 122, 134, 136–139,
141, 143, 149, 151–153, 156–158, 160, 161,
172–177
-3 tensor, 12–14, 19, 54, 60, 66, 72, 77–84, 86–
88, 99, 107, 112
-4 tensor, 65, 66, 72, 79, 107, 167
of tensor, 9, 12, 14, 16, 19, 20, 42, 50–54, 58,
59, 61–70, 72, 73, 77–79, 84, 88, 93, 97,
107, 120, 123, 128, 130–132, 135, 137–143,
146, 152, 157, 160, 161
Rate of
deformation tensor, 176, 179
strain tensor, 7, 176, 179
Real, 9, 11, 24, 27, 29, 35, 44, 103
Reciprocal, 30, 31, 39, 41, 43, 44, 53, 63, 91, 92, 95
metric tensor, 91
Reciprocity relation, 53, 54, 72, 91
Rectangular
Cartesian, 25, 26, 29, 33, 34, 54, 56, 72, 92
coordinate system, 25, 26, 29, 33, 34, 54, 56,
62, 72, 92, 112, 129, 132, 146
parallelepiped, 105
Rectilinear coordinate system, 24–26, 30, 34, 35, 47,
121, 123–125, 129, 132, 136
Reference frame, 32
Reflection, 33, 34, 55
Relative
permutation tensor, 8, 81, 99, 107, 110
scalar, 94, 128, 136
tensor, 7, 57, 58, 63, 66, 72, 78, 81, 89, 107,
128, 136
Replacement operator, 52, 82, 88, 93, 107, 126, 136
Rheology, 175
Ricci
curvature scalar, 7, 172, 173, 179
curvature tensor, 7, 172, 173, 178, 179
theorem, 114, 115, 124–126, 135
Riemann-Christoffel curvature tensor, 7, 22, 23, 127,
128, 167–172, 178, 179
Riemannian
curvature, 23
geometry, 22, 24, 46
metric, 22, 94
space, 22, 23, 46, 93, 170
Right handed system, 41, 42, 49, 84
Rotation, 32–34, 58, 175, 176
Scalar, 6, 9–13, 16–20, 31, 34, 58, 59, 62–66, 68–70,
72, 73, 94, 103, 123, 124, 128–132, 135–
145, 147–151, 153, 157
field, 19, 50, 129, 136, 152–154, 158
invariant, 155, 157, 158, 177
multiplication, 34
operator, 137, 138, 152
triple product, 41, 84, 100, 101, 105, 106, 110,
111, 157, 177
Scale factor, 7, 36, 39, 41, 48, 69–71, 95, 100, 103,
105, 110, 111, 117–119, 146, 148, 150, 153
Schur theorem, 23, 46
Semicircle, 26, 36
plane, 26, 35
Semicolon notation, 6, 10, 11, 19, 124
Shear stress, 174
Shifting operator, 40, 52, 93, 98, 109, 113, 114, 126,
135, 136
Skew-symmetric, 59, 73, 169
Solenoidal, 160
Sphere, 23, 25, 26, 35, 36, 46
Spherical coordinate system, 6, 7, 10, 24–26, 28, 35,
38, 39, 47, 48, 71, 75, 95, 110, 111, 113,
117–119, 134, 137, 150, 152–154, 175
Stokes theorem, 155, 160, 161, 177
Straight line, 23, 25, 26, 35, 36, 47, 55
Stress tensor, 8, 174, 179
Subtraction of tensors, 62, 63, 73
Sum rule, 132
Summation, 11, 14–16, 20, 43, 62, 66, 113, 147
convention, 11, 14, 15, 19, 83, 87, 137, 148, 150
Surface, 6, 7, 12, 23–25, 35, 46, 53, 76, 94, 103–105,
119, 160, 161, 174
Symbolic notation, 9, 12, 19, 65, 67, 124, 126, 131,
135, 139, 156
Symmetric, 18, 22, 50, 59–62, 72, 73, 76, 77, 84,
91–93, 109, 113, 116, 118, 133, 169–172,
174–176, 178, 179
tensor, 59–62, 73, 84, 172, 176
Symmetry, 59, 61, 62, 73, 84, 107, 108, 119, 144,
153, 169, 173, 178, 179
Tangent vector, 32, 39, 43, 55, 98, 103, 129, 131
Tensor
algebra, 42, 62, 63, 156
calculus, 1, 11, 12, 14, 17, 19, 22, 24, 57, 62, 65,
67, 69, 75–77, 86, 91, 112, 121, 137, 152,
155, 157, 161, 167, 168, 173, 177
component, 18, 20, 55, 78, 86, 112, 123, 132,
143, 152
equality, 14–18, 20, 58, 62, 63, 72, 73, 77
185
expression, 14–18, 20, 56, 58, 63, 72, 87
field, 19, 21, 136, 152, 160, 161
identity, 76, 87, 89, 155, 161
multiplication, 64, 65, 73, 124, 131
notation, 12, 13, 17, 71, 84, 92, 99, 109, 110,
137, 152, 153, 155, 157, 158, 160, 161, 173–
177, 179
representation, 50, 69–71, 74
term, 14–18, 20, 56–58, 60, 62, 63, 72, 82, 84,
86, 120, 122–124, 126, 128, 132, 134–136,
144, 162, 164, 172
test, 69, 74
Torus, 23
Total
derivative, 112, 130, 132, 133, 136
differentiation, 112, 130, 132, 136
stress tensor, 174
Totally
anti-symmetric, 61, 72, 78, 81, 84, 107
covariant Riemann-Christoffel curvature tensor,
167
symmetric, 61, 72
Trace, 7, 66, 83, 139, 156, 158, 177
Traction vector, 7, 174
Transformation, 7, 11–13, 15, 19, 22, 26, 27, 29–34,
36, 40–42, 47–51, 55, 57, 58, 61, 62, 69,
71–74, 76–78, 81, 91, 93, 94, 98, 103, 105,
107, 109, 112, 121, 141, 153, 157
Translation, 33, 34, 175
Triad, 54, 65
True
scalar, 31, 157
tensor, 55–58, 63, 72
vector, 31, 55, 57, 72
notation, 12, 158, 160, 177
operator, 137, 138, 140, 152
triple product, 101, 102, 110, 157, 177
Velocity
gradient tensor, 6, 175, 176, 179
vector, 7, 175
Volume, 6, 7, 39, 41, 105, 106, 111, 160
Vorticity tensor, 7, 8, 176, 179
Weight of tensor, 7, 53, 57, 58, 63, 64, 66, 72, 73,
78, 81, 89, 94, 107, 128, 136
Zero
Unit
tensor, 14
vector, 55, 56, 84, 160
Unity
matrix, 53
tensor, 14, 92, 94, 109, 175
Vector, 6, 7, 9, 11–13, 15–17, 19, 20, 25–27, 31, 32,
39–44, 48, 53–59, 64–74, 83, 84, 87, 91–94,
97–103, 105, 108, 110, 111, 113, 121–124,
127–131, 134, 137–144, 146, 147, 149–153,
156–158, 160, 163, 164, 168, 174, 175, 177
algebra, 155
calculus, 74, 137, 140, 147, 148, 152, 155, 157,
158
field, 19, 138, 152–154, 158, 160, 177, 178
identity, 87, 161, 177
186
tensor, 14, 58, 62, 72, 73, 133
vector, 58
Author Notes
• All copyrights of this book are held by the author.
• This book, like any other academic document, is protected by the terms and conditions
of the universally recognized intellectual property rights. Hence, any quotation or use of
any part of the book should be acknowledged and cited according to the scholarly approved
traditions.
187
About the Book
This book is an intermediate level text on tensor calculus.
The formulation is largely based on assuming a general curvilinear coordinate system. The book is furnished with an index and thorough sets of exercises. Cross referencing, which
is hyperlinked in the electronic version, is used extensively
throughout the book to facilitate linking related topics. The
book also contains a number of graphic illustrations to visualize the abstract ideas and consolidate the subtle concepts.
The book can be used as a textbook or as a reference for an
introductory or intermediate level course on tensor calculus
or related mathematical and scientific subjects.
About the Author
The author of the book possesses a diverse academic and research background. He holds a BEng in electronics engineering, a BSc in physics, a PhD in petroleum engineering, a PhD
in crystallography, and a PhD in atomic physics and astronomy. He also worked as a postdoctoral research associate for
several years. He published dozens of scientific research papers in many refereed journals and produced numerous academically oriented documents which are freely available on
the world wide web.