Academia.eduAcademia.edu

Principles of Tensor Calculus

2017, Principles of Tensor Calculus

This book is based on my previous book: Tensor Calculus Made Simple, where the development of tensor calculus concepts and techniques are continued at a higher level. Unlike the previous book which is largely based on a Cartesian approach, the formulation in the present book is based on a general coordinate system. The book is furnished with an index as well as detailed sets of exercises to provide useful revision and practice. To facilitate linking related concepts and sections, cross referencing is used extensively throughout the book. The book also contains a number of graphic illustrations to help the readers to visualize the ideas and understand the subtle concepts. The book can be used as a text for an introductory or an intermediate level course on tensor calculus.

Principles of Tensor Calculus gi j = Ei · E j ∂r ∂ ui u1 CS ∇2 f= δ Ak δt √1 g ∂i = E2 j E1 i E1 u2 CC Aj E2 =A i E3 u2 CS j1 ... jn u1 CC √ gg i j ∂j f  dAk dt δ jin1 δ jin2 .. . δ jinn ··· ··· .. . ··· E3 ...in δ ji11 ... jn = ε i1 ...in ε δ ji21 δ ji22 .. . δ ji2n δ Ei = ...in δ ji11 ... jn = u3 CC δ ji11 δ ji12 .. . δ ji1n du j i k + Γi j A dt u3 CS i j g i E i ∇u = = i j E ·E Taha Sochi Preface This book is based on my previous book: Tensor Calculus Made Simple, where the development of tensor calculus concepts and techniques are continued at a higher level. In the present book, we continue the discussion of the main topics of the subject at a more advanced level expanding, when necessary, some topics and developing further concepts and techniques. The purpose of the present book is to solidify, generalize, fill the gaps and make more rigorous what have been presented in the previous book. Unlike the previous book which is largely based on a Cartesian approach, the formulation in the present book is largely based on assuming an underlying general coordinate system although some example sections are still based on a Cartesian approach for the sake of simplicity and clarity. The reader will be notified about the underlying system in the given formulation. We also provide a sample of formal proofs to familiarize the reader with the tensor techniques. However, due to the preset objectives and the intended size of the book, we do not offer comprehensive proofs and complete theoretical foundations for the provided materials although we generally try to justify many of the given formulations descriptively or by interlinking to related formulations or by similar pedagogical techniques. This may be seen as a more friendly method for constructing and establishing the abstract concepts and techniques of tensor calculus. The book is furnished with an index in the end of the book as well as rather detailed sets of exercises in the end of each chapter to provide useful revision and practice. To facilitate linking related concepts and sections, and hence ensure better understanding of the given materials, cross referencing, which is hyperlinked for the ebook users, is used extensively throughout the book. The book also contains a number of graphic illustrations to help the readers to visualize the ideas and understand the subtle concepts. The book can be used as a text for an introductory or an intermediate level course on tensor calculus. The familiarity with the materials presented in the previous book will be an advantage although it is not necessary for someone with a reasonable mathematical background. Moreover, the main materials of the previous book are absorbed within the structure of the present book for the sake of completeness and to make the book rather self-contained considering the predetermined objectives. I hope I achieved these goals. Taha Sochi London, August 2017 1 Contents Preface 1 Table of Contents 2 Nomenclature 6 1 Preliminaries 1.1 General Conventions and Notations . . . . . . . . . . . . . . . . . . . . . . 1.2 General Background about Tensors . . . . . . . . . . . . . . . . . . . . . . 1.3 Exercises and Revision . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9 9 12 19 2 Spaces, Coordinate Systems and Transformations 2.1 Spaces . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2 Coordinate Systems . . . . . . . . . . . . . . . . . . . . . . . 2.2.1 Rectilinear and Curvilinear Coordinate Systems . . . 2.2.2 Orthogonal Coordinate Systems . . . . . . . . . . . . 2.2.3 Homogeneous Coordinate Systems . . . . . . . . . . . 2.3 Transformations . . . . . . . . . . . . . . . . . . . . . . . . . 2.3.1 Proper and Improper Transformations . . . . . . . . 2.3.2 Active and Passive Transformations . . . . . . . . . . 2.3.3 Orthogonal Transformations . . . . . . . . . . . . . . 2.3.4 Linear and Nonlinear Transformations . . . . . . . . 2.4 Coordinate Curves and Coordinate Surfaces . . . . . . . . . 2.5 Scale Factors . . . . . . . . . . . . . . . . . . . . . . . . . . 2.6 Basis Vectors and Their Relation to the Metric and Jacobian 2.7 Relationship between Space, Coordinates and Metric . . . . 2.8 Exercises and Revision . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22 22 24 25 26 27 29 32 33 33 34 35 36 39 44 46 3 Tensors 3.1 Tensor Types . . . . . . . . . . . . . . . . . . . 3.1.1 Covariant and Contravariant Tensors . . 3.1.2 True and Pseudo Tensors . . . . . . . . . 3.1.3 Absolute and Relative Tensors . . . . . . 3.1.4 Isotropic and Anisotropic Tensors . . . . 3.1.5 Symmetric and Anti-symmetric Tensors . 3.1.6 General and Affine Tensors . . . . . . . . 3.2 Tensor Operations . . . . . . . . . . . . . . . . 3.2.1 Addition and Subtraction . . . . . . . . 3.2.2 Multiplication of Tensor by Scalar . . . . 3.2.3 Tensor Multiplication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50 50 50 55 57 58 59 62 62 63 64 64 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66 66 68 69 69 71 4 Special Tensors 4.1 Kronecker delta Tensor . . . . . . . . . . . . . . . . . . . . . . . 4.2 Permutation Tensor . . . . . . . . . . . . . . . . . . . . . . . . . 4.3 Identities Involving Kronecker or/and Permutation Tensors . . . 4.3.1 Identities Involving Kronecker delta Tensor . . . . . . . . 4.3.2 Identities Involving Permutation Tensor . . . . . . . . . 4.3.3 Identities Involving Kronecker and Permutation Tensors 4.4 Generalized Kronecker delta Tensor . . . . . . . . . . . . . . . . 4.5 Metric Tensor . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.6 Definitions Involving Special Tensors . . . . . . . . . . . . . . . 4.6.1 Dot Product . . . . . . . . . . . . . . . . . . . . . . . . . 4.6.2 Magnitude of Vector . . . . . . . . . . . . . . . . . . . . 4.6.3 Angle between Vectors . . . . . . . . . . . . . . . . . . . 4.6.4 Cross Product . . . . . . . . . . . . . . . . . . . . . . . . 4.6.5 Scalar Triple Product . . . . . . . . . . . . . . . . . . . . 4.6.6 Vector Triple Product . . . . . . . . . . . . . . . . . . . 4.6.7 Determinant of Matrix . . . . . . . . . . . . . . . . . . . 4.6.8 Length . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.6.9 Area . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.6.10 Volume . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.7 Exercises and Revision . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76 76 77 82 82 83 85 88 91 97 97 98 98 99 100 101 102 102 103 105 107 5 Tensor Differentiation 5.1 Christoffel Symbols . . . 5.2 Covariant Differentiation 5.3 Absolute Differentiation 5.4 Exercises and Revision . 3.3 3.4 3.2.4 Contraction . . 3.2.5 Inner Product . 3.2.6 Permutation . . 3.2.7 Tensor Test and Tensor Representations Exercises and Revision . . . . . . . . . . . . . . . . . . . . . . . . . . . Quotient Rule . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 112 112 120 129 133 6 Differential Operations 6.1 Cartesian Coordinate System 6.1.1 Operators . . . . . . . 6.1.2 Gradient . . . . . . . . 6.1.3 Divergence . . . . . . . 6.1.4 Curl . . . . . . . . . . 6.1.5 Laplacian . . . . . . . 6.2 General Coordinate System . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 137 138 138 138 139 139 140 140 . . . . . . . . 3 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 140 140 142 143 144 146 146 147 147 147 148 148 148 149 149 149 149 150 150 150 151 151 151 152 7 Tensors in Application 7.1 Tensors in Mathematics . . . . . . . . . . . . . . . . . . . . . . 7.1.1 Common Definitions in Tensor Notation . . . . . . . . . 7.1.2 Scalar Invariants of Tensors . . . . . . . . . . . . . . . . 7.1.3 Common Identities in Vector and Tensor Notation . . . . 7.1.4 Integral Theorems in Tensor Notation . . . . . . . . . . . 7.1.5 Examples of Using Tensor Techniques to Prove Identities 7.2 Tensors in Geometry . . . . . . . . . . . . . . . . . . . . . . . . 7.2.1 Riemann-Christoffel Curvature Tensor . . . . . . . . . . 7.2.2 Bianchi Identities . . . . . . . . . . . . . . . . . . . . . . 7.2.3 Ricci Curvature Tensor and Scalar . . . . . . . . . . . . 7.3 Tensors in Science . . . . . . . . . . . . . . . . . . . . . . . . . . 7.3.1 Infinitesimal Strain Tensor . . . . . . . . . . . . . . . . . 7.3.2 Stress Tensor . . . . . . . . . . . . . . . . . . . . . . . . 7.3.3 Displacement Gradient Tensors . . . . . . . . . . . . . . 7.3.4 Finger Strain Tensor . . . . . . . . . . . . . . . . . . . . 7.3.5 Cauchy Strain Tensor . . . . . . . . . . . . . . . . . . . . 7.3.6 Velocity Gradient Tensor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 155 155 156 157 158 160 161 167 167 171 172 173 173 174 174 175 175 175 6.3 6.4 6.5 6.6 6.2.1 Operators . . . . . . . . 6.2.2 Gradient . . . . . . . . . 6.2.3 Divergence . . . . . . . . 6.2.4 Curl . . . . . . . . . . . 6.2.5 Laplacian . . . . . . . . Orthogonal Coordinate System 6.3.1 Operators . . . . . . . . 6.3.2 Gradient . . . . . . . . . 6.3.3 Divergence . . . . . . . 6.3.4 Curl . . . . . . . . . . . 6.3.5 Laplacian . . . . . . . . Cylindrical Coordinate System . 6.4.1 Operators . . . . . . . . 6.4.2 Gradient . . . . . . . . . 6.4.3 Divergence . . . . . . . . 6.4.4 Curl . . . . . . . . . . . 6.4.5 Laplacian . . . . . . . . Spherical Coordinate System . . 6.5.1 Operators . . . . . . . . 6.5.2 Gradient . . . . . . . . . 6.5.3 Divergence . . . . . . . . 6.5.4 Curl . . . . . . . . . . . 6.5.5 Laplacian . . . . . . . . Exercises and Revision . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.4 7.3.7 Rate of Strain Tensor . . . . . . . . . . . . . . . . . . . . . . . . . . 176 7.3.8 Vorticity Tensor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 176 Exercises and Revision . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 177 References 180 Index 181 Author Notes 187 5 Nomenclature In the following list, we define the common symbols, notations and abbreviations which are used in the book as a quick reference for the reader. ∇ ∇; and ∇; ∇f ∇·A ∇×A ∇2 , ∂ii , ∇ii ∇v, ∂i vj , (subscript) ; (subscript) hat (e.g. Âi , Êi ) bar (e.g. ūi , Āi ) ◦ ⊥ 1D, 2D, 3D, nD δ/δt ∂i and ∇i ∂;i [ij, k] A B, Bij B−1 , Bij−1 C Cn d, di det dr ds dσ dτ ei er , eθ , eφ err , erθ , · · · , eφφ eρ , eφ , ez eρρ , eρφ , · · · , ezz E, Eij Ei , Ei Ei nabla differential operator covariant and contravariant differential operators gradient of scalar f divergence of tensor A curl of tensor A Laplacian operator velocity gradient tensor partial derivative with respect to following index(es) covariant derivative with respect to following index(es) physical representation or normalized vector transformed quantity inner or outer product operator perpendicular to one-, two-, three-, n-dimensional absolute derivative operator with respect to t partial derivative operator with respect to ith variable covariant derivative operator with respect to ith variable Christoffel symbol of 1st kind area Finger strain tensor Cauchy strain tensor curve of class n displacement vector determinant of matrix differential of position vector length of infinitesimal element of curve area of infinitesimal element of surface volume of infinitesimal element of space ith vector of orthonormal vector set (usually Cartesian basis set) basis vectors of spherical coordinate system unit dyads of spherical coordinate system basis vectors of cylindrical coordinate system unit dyads of cylindrical coordinate system first displacement gradient tensor ith covariant and contravariant basis vectors ith orthonormalized covariant basis vector 6 Eq./Eqs. g g gij , g ij , gij g11 , g12 , · · · gnn g 11 , g 12 , · · · g nn hi iff J J J−1 L n, ni P P (n, k) qi qi r R Rij , Rij Rijkl , Rijkl r, θ, φ S S, Sij S̄, S̄ij t T (superscript) T, Ti tr ui v, vi V w xi , xi ′ xi , xi x, y, z γ, γij γ̇ Γkij δ δij , δ ij , δij ijk ij n , δji11 ...i , δlmn δkl ...jn ∆, ∆ij Equation/Equations determinant of covariant metric tensor metric tensor covariant, contravariant and mixed metric tensor or its components coefficients of covariant metric tensor coefficients of contravariant metric tensor scale factor for ith coordinate if and only if Jacobian of transformation between two coordinate systems Jacobian matrix of transformation between two coordinate systems inverse Jacobian matrix of transformation length of curve normal vector to surface point k-permutations of n objects ith coordinate of orthogonal coordinate system ith unit basis vector of orthogonal coordinate system position vector Ricci curvature scalar Ricci curvature tensor of 1st and 2nd kind Riemann-Christoffel curvature tensor of 1st and 2nd kind coordinates of spherical coordinate system surface rate of strain tensor vorticity tensor time transposition of matrix traction vector trace of matrix ith coordinate of general coordinate system velocity vector volume weight of relative tensor ith Cartesian coordinate ith Cartesian coordinate of particle at past and present times coordinates of 3D space (mainly Cartesian) infinitesimal strain tensor rate of strain tensor Christoffel symbol of 2nd kind Kronecker delta tensor covariant, contravariant and mixed ordinary Kronecker delta generalized Kronecker delta in 2D, 3D and nD space second displacement gradient tensor 7 ǫij , ǫijk , ǫi1 ...in ǫij , ǫijk , ǫi1 ...in ǫij , ǫijk , ǫi1 ...in ǫij , ǫijk ,ǫi1 ...in ρ, φ ρ, φ, z σ, σij ω Ω covariant relative permutation tensor in 2D, 3D and nD space contravariant relative permutation tensor in 2D, 3D and nD space covariant absolute permutation tensor in 2D, 3D and nD space contravariant absolute permutation tensor in 2D, 3D and nD space coordinates of plane polar coordinate system coordinates of cylindrical coordinate system stress tensor vorticity tensor region of space 8 Chapter 1 Preliminaries In this introductory chapter, we provide preliminary materials about conventions and notations as well as basic facts about tensors which will be needed in the subsequent parts of the book. The chapter is therefore divided into two sections: the first is about general conventions and notations used in the book, and the second is on general background about tensors. 1.1 General Conventions and Notations In this section, we provide general notes about the main conventions and notations used in the present book. We usually use the term “tensor” to mean tensors of all ranks including scalars (rank-0) and vectors (rank-1). However, we may also use this term as opposite to scalar and vector, i.e. tensor of rank-n where n > 1. In almost all cases, the meaning should be obvious from the context. We note that in the present book all tensors of all ranks and types are assumed to be real quantities, i.e. they have real rather than imaginary or complex components. We use non-indexed lower case light face italic Latin letters (e.g. f and h) to label scalars, while we use non-indexed lower or upper case bold face non-italic Latin letters (e.g. a and A) to label vectors in symbolic notation. The exception to this is the basis vectors where indexed bold face lower or upper case non-italic symbols (e.g. e1 and Ei ) are used. However, there should be no confusion or ambiguity about the meaning of any one of these symbols. We also use non-indexed upper case bold face non-italic Latin letters (e.g. A and B) to label tensors of rank > 1 in symbolic notation. Since matrices in this book are supposed to represent rank-2 tensors, they also follow the rules of labeling tensors symbolically by using non-indexed upper case bold face non-italic Latin letters. We note that in a few cases in the final chapter (see § 7.3) we used boldface and indexed light face Greek symbols to represent particular tensors, which are commonly labeled in the literature by these symbols, to keep with the tradition. Indexed light face italic Latin symbols (e.g. ai and Bijk ) are used in this book to denote tensors of rank > 0 in their explicit tensor form, i.e. index notation. Such symbols may also be used to denote the components of these tensors. The meaning is usually transparent and can be identified from the context if it is not declared explicitly. Tensor indices in this book are lower case Latin letters which may be taken preferably from the middle of the Latin alphabet (such as i, j and k) for the free indices and from the beginning of the Latin alphabet (such as a and b) for the dummy indices. We also use numbered indices, such as (i1 , i2 , . . . , ik ), for this purpose when the number of tensor indices is variable. Numbers are also used as indices in some occasions (e.g. ǫ12 ) for obvious purposes such as making statements about particular components. 9 1.1 General Conventions and Notations 10 Partial derivative symbol with a subscript index (e.g. ∂i ) is used to denote partial differentiation with respect to the ith variable, that is: ∂i = ∂ ∂xi (1) However,we should note that in this book we generalize partial derivative notation so that ∂i symbolizes partial derivative with respect to the ui coordinate of general coordinate systems and not just Cartesian coordinates which are usually denoted by xi or xi . The type of coordinates, being Cartesian or general or otherwise, will be determined by the context which should be obvious in all cases. Similarly, we use partial derivative symbol with a twice-repeated index to denote the Laplacian operator, that is:[1] ∂ii = ∂i ∂i = ∇2 (2) Partial derivative symbol with a coordinate label subscript, rather than an index, is also used to denote partial differentiation with respect to that spatial variable. For instance: ∂r = ∂ ∂r (3) is used to denote the partial derivative with respect to the radial coordinate r in spherical coordinate systems which are identified by the spatial variables (r, θ, φ). It should be obvious that in notations like ∂r the subscript is used as a label rather than an index and hence it does not follow the rules of tensor indices which will be discussed later (see § 1.2). Following the widely used convention, a subscript comma preceding a subscript index (e.g. Ak,i ) is used to denote partial differentiation with respect to the spatial coordinate which is indexed by the symbol that follows the comma. For example, f,i and Ajk,i are used to represent the partial derivative of the scalar f and rank-2 tensor Ajk with respect to the ith coordinate, that is: f,i = ∂i f Ajk,i = ∂i Ajk (4) We also follow the common convention of using a subscript semicolon preceding a subscript index (e.g. Akl;i ) to symbolize the operation of covariant differentiation with respect to the ith coordinate (see § 5.2). The semicolon notation may also be attached to the normal differential operators for the same purpose. For example, ∇;i and ∂;i symbolize covariant differential operators with respect to the ith variable. In this regard, we should remark that more than one index may follow the comma and semicolon in these notations to represent multiple partial and covariant differentiation with respect to the indexed variables according to the stated order of the indices. For example, Ai,jk is used to represent the mixed second order partial derivative of the tensor Ai with respect to the j th and k th coordinates, while Bji;km is used to represent the mixed second [1] This is the Cartesian form. For the other forms, the reader is referred to § 6. 1.1 General Conventions and Notations 11 order covariant derivative of the tensor Bji with respect to the k th and mth coordinates, that is:  Bji;km = ∇;m (∇;k Bji ) (5) Ai,jk = ∂k ∂j Ai We also note that in a few occasions superscripts, rather than subscripts, comma and semicolon preceding superscript index (e.g. f ,i and Ai;j ) are used to represent contravariant partial derivative and contravariant tensor derivative respectively. Superscripted differential operators (e.g. ∂ i and ∇;i ) are also used occasionally to represent these differential operators in their contravariant form. A matter related to tensor differentiation is that we follow the conventional notation δtδ to represent the intrinsic derivative, which is also known as the absolute derivative, with respect to the variable t along a given curve, as will be discussed in § 5.3. Due to the restriction that we impose of using real (as opposite to imaginary and complex) quantities exclusively in this book, all arguments of real-valued functions which are not defined for negative quantities, like square roots and logarithmic functions, are assumed to be non-negative by taking the absolute value, if necessary, without using the absolute value symbol. This is to simplify p the notation and avoid potential confusion with √ the determinant notation. So, g means |g| and ln(g) means ln(|g|). We follow the summation convention which is widely used in the literature of tensor calculus and its applications. However, the summation symbol (i.e. Σ) is used in a few cases where a summation operation is needed but the conditions of the summation convention do not apply or there is an ambiguity about them, e.g. when an index is repeated more than twice or when a summation index is not repeated visually because it is part of a squared symbol. In a few other cases, where a twice-repeated index that complies with the conditions of the summation convention does not imply summation and hence the summation convention do not apply, we clarified the situation by adding comments like “no sum on index”. We may also add a “no sum” comment in some cases where the conditions of the summation convention do not apply technically but the expression may be misleading since it contains a repetitive index, e.g. when both indices are of the same variance type in a general coordinate system such as g ii or when one of the apparent indices is in fact a label for a scalar rather than a variable index such as |Ei | or hj . All the transformation equations in the present book are continuous and real, and all the derivatives are continuous over their intended domain. Based on the well known continuity condition of differential calculus, the individual differential operators in the second (and higher) order partial derivatives with respect to different indices are commutative, that is: ∂i ∂j = ∂j ∂i (6) We generally assume that this continuity condition is satisfied and hence the order of the partial differential operators in these mixed second order partial derivatives does not matter. We use vertical bars (i.e. |::|) to symbolize determinants and square brackets (i.e. [::]) to symbolize matrices. This applies when these symbols contain arrays of objects; otherwise they have their normal meaning according to the context, e.g. bars embracing a vector 1.2 General Background about Tensors 12 such as |v| mean modulus of the vector. Also, we use indexed square brackets (such as [A]i and [∇f ]i ) to denote the ith component of vectors in their symbolic or vector notation. For tensors of higher rank, more than one index are used to denote their components, e.g. [A]ij represents the ij th component of the rank-2 tensor A. We finally should remark that although we generally talk about nD spaces, our main focus is the low dimensionality spaces (mostly 2D and 3D) especially with regard to coordinate systems and hence some of the statements may apply only to these low dimensionality spaces although the statements are given in the context of nD spaces. In most cases, such statements can be generalized simply by adding extra conditions or by a slight modification to the phrasing and terminology. 1.2 General Background about Tensors A tensor is an array of mathematical objects (usually numbers or functions) which transforms according to certain rules under coordinates change. In an nD space, a tensor of rank-k has nk components which may be specified with reference to a given coordinate system. Accordingly, a scalar, such as temperature, is a rank-0 tensor with (assuming a 3D space) 30 = 1 component, a vector, such as force, is a rank-1 tensor with 31 = 3 components, and stress is a rank-2 tensor with 32 = 9 components. In Figure 1 we graphically illustrate the structure of a rank-3 tensor in a 3D space. The nk components of a rank-k tensor in an nD space are identified by k distinct integer indices (e.g. i, j, k) which are attached, according to the commonly-employed tensor notation, as superscripts or subscripts or a mix of these to the right side of the symbol utilized to label the tensor, e.g. Aijk , Aijk and Ajk i . Each tensor index takes all the values over a predefined range of dimensions such as 1 to n in the above example of an nD space. In general, all tensor indices have the same range, i.e. they are uniformly dimensioned.[2] When the range of tensor indices is not stated explicitly, it is usually assumed to range over the values 1, 2, 3. However, the range must be stated explicitly or implicitly to avoid ambiguity. The characteristic property of tensors is that they satisfy the principle of invariance under certain coordinate transformations. Therefore, formulating the fundamental laws of physics in a tensor form ensures that they are form-invariant, and hence they are objectively representing the physical reality and do not depend on the observer and his coordinate system. Having the same form in different coordinate systems may also be labeled as being covariant although this term is usually used for a different meaning in tensor calculus, as will be explained in § 3.1.1. While tensors of rank-0 are generally represented in a common form of light face nonindexed italic symbols like f and h, tensors of rank ≥ 1 are represented in several forms and notations, the main ones are the index-free notation, which may also be called the direct or symbolic or Gibbs notation, and the indicial notation which is also called the [2] This applies to the common cases of tensor applications, but there are instances (e.g. in differential geometry of curves and surfaces) of tensors which are not uniformly dimensioned because the tensor is related to two spaces with different dimensions such as a 2D surface embedded in a 3D space. 1.2 General Background about Tensors 13 A332 A331 A321 A311 A232 A231 A221 A222 A211 A131 A121 A111 A132 A122 A112 A212 A322 A312 A333 A323 A313 A233 A223 A213 A133 A123 A113 Figure 1: Graphical illustration of a rank-3 tensor Aijk in a 3D space, i.e. each one of i, j, k ranges over 1, 2, 3. index or component or tensor notation. The first is a geometrically oriented notation with no reference to a particular coordinate system and hence it is intrinsically invariant to the choice of coordinate systems, whereas the second takes an algebraic form based on components identified by indices and hence the notation is suggestive of an underlying coordinate system, although being a tensor makes it form-invariant under certain coordinate transformations and therefore it possesses certain invariant properties. The index-free notation is usually identified by using bold face non-italic symbols, like a and B, while the indicial notation is identified by using light face indexed italic symbols such as ai and Bij . It is noteworthy that although rank-0 and rank-1 tensors are, respectively, scalars and vectors, not all scalars and vectors (in their generic sense) are tensors of these ranks. Similarly, rank-2 tensors are normally represented by square matrices but not all square matrices represent rank-2 tensors. 1.2 General Background about Tensors 14 Tensors can be combined through common algebraic operations such as addition and multiplication. Tensor term is a product of tensors including scalars and vectors and may consist of a single tensor which can be regarded as a multiple of unity. Tensor expression is an algebraic sum of tensor terms which may be a trivial sum in the case of a single term. Tensor equality is an equality of two tensor terms and/or expressions. An index that occurs once in a tensor term is a free index while an index that occurs twice in a tensor term is a dummy or bound index. The order of a tensor is identified by the number of its indices. For example, Aijk is a tensor of order 3 while Bkm is a tensor of order 2. The order of the tensor normally identifies its rank as well and hence Aijk is of rank-3 and Bkm is of rank-2. However, when the operation of contraction of indices (see § 3.2.4) takes place once or more, the order of the tensor is not affected but its rank is reduced by two for each contraction operation. Hence, the order of a tensor is equal to the number of all of its indices including the dummy indices, while the rank is equal to the number of its free indices only. Accordingly, Aab abmn a is of order 6 and rank-2 while Bai is of order 3 and rank-1. We note that many authors follow different conventions such as using “order” as equivalent to what we call “rank”. Tensors whose all indices are subscripts, like Aij , are called covariant, while tensors whose all indices are superscripts, like Ak , are called contravariant. Tensors with both types of indices, like Aamn ak , are called mixed type. Subscript indices, rather than subscripted tensors, are also described as covariant and superscript indices are described as contravariant. The zero tensor is a tensor whose all components are zero. The unit tensor or unity tensor, which is usually defined for rank-2 tensors, is a tensor whose all elements are zero except those with identical values of all indices (e.g. A11 or B 33 ) which are assigned the value 1. There are general rules that govern the manipulation of tensors and hence they should be observed in the handling of mathematical expressions and calculations of tensor calculus. One of these rules is that no tensor index is allowed to occur more than twice in a legitimate tensor term. However, we follow in this assertion the common literature of tensor calculus which represents the ordinary use of repeated indices in tensor terms. In fact, there are many instances in the literature of tensor calculus where indices are legitimately repeated more than twice in a single term. The bottom line is that as long as the tensor expression makes sense and the intention is clear, such repetitions should be allowed with no need to take special precautions like using parentheses as done by some authors. In particular, the forthcoming summation convention will not apply automatically in such cases although summation on such indices, if needed, can be carried out explicitly, by using the summation symbol Σ, or by a special declaration of such intention similar to the summation convention which is usually restricted to the twice-repeated indices. Regarding the aforementioned summation convention, according to this convention which is widely used in the literature of tensor calculus including the present book, dummy indices imply summation over their range. More clearly, a twice-repeated variable (i.e. not numeric) index in a single term implies a sum of terms equal in number to the range of the repeated index. Hence, in a 3D space we have: 1j 2j 3j Aaj a = A1 + A2 + A3 (7) 1.2 General Background about Tensors 15 while in a 4D space we have: Bi + Cia Da = Bi + Ci1 D1 + Ci2 D2 + Ci3 D3 + Ci4 D4 (8) We note that although the twice-repeated index should be in the same term for the summation convention to apply, it does not matter if the two indices occur in one tensor or in two tensors, as seen in the last example where the summation convention applies to a. We should also remark that there are many cases in the mathematics of tensors where a repeated index is needed but with no intention of summation, such as using Aii or Ajj to mean the components of the tensor A whose indices have identical value like A11 and A22 . So to avoid confusion, when the dummy indices in a particular case do not imply summation, the situation must be clarified by enclosing such indices in parentheses or by underscoring or by using upper case letters with declaration of these conventions, or by adding a clarifying comment like “no summation over repeated indices”. These precautions are obviously needed if the summation convention is adopted in general but it does not apply in some exceptional cases where repeated indices are needed in the notation with no implication of summation. Another rule of tensors is that a free index should be understood to vary over its range (e.g. 1, . . . , n) which is determined by the space dimension and hence it should be interpreted as saying “for all components represented by the index”. Therefore, a free index represents a number of terms or expressions or equalities equal to the number of the allowed values of its range. For example, when i and j can vary over the range 1, . . . , n then the expression Ai + Bi represents n separate expressions while the equation Aji = Bij represents n × n separate equations which represent the combination of all possible n values of i with all possible n values of j, that is: A1 + B1 , A2 + B2 , · · · , An + Bn (9) A11 = B11 , A21 = B12 , A12 = B21 , · · · , Ann = Bnn (10) Also, each tensor index should conform to one of the forthcoming variance transformation rules as given by Eqs. 71 and 72, i.e. it is either covariant or contravariant. For orthonormal Cartesian coordinate systems, the two variance types (i.e. covariant and contravariant) do not differ because the metric tensor is given by the Kronecker delta (refer to § 4.1 and 4.5) and hence any index can be upper or lower although it is common to use lower indices in this case. We note that orthonormal vectors mean a set of vectors which are mutually orthogonal and each one is of unit length, while orthonormal coordinate system means a coordinate system whose basis vector set is orthonormal at all points of the space where the system is defined (see § 2.6). The orthonormality of vectors may be expressed mathematically by: Vi · Vj = δij or Vi · Vj = δ ij (11) where the indexed δ is the Kronecker delta symbol and the indexed V symbolizes a vector in the set. 1.2 General Background about Tensors 16 In this context, we should remark that for tensor invariance, a pair of dummy indices involved in summation should in general be complementary in their variance type, i.e. one covariant and the other contravariant. However, for orthonormal Cartesian systems the two variance types are the same and hence when both dummy indices are covariant or both are contravariant it should be understood as an indication that the underlying coordinate system is orthonormal Cartesian if the possibility of an error is excluded. As indicated earlier, tensor order is equal to the number of its indices while tensor rank is equal to the number of its free indices. Hence, scalars (terms, expressions and equalities) have no free index since they are of rank-0, and vectors have a single free index while rank-2 tensors have exactly two free indices. Similarly, rank-n tensors have exactly n free indices. The dimension of a tensor is determined by the range taken by its indices which represents the number of dimensions of the underlying space. For example, in a 3D space the tensor Aji is of rank-2 because it possesses exactly two free indices but it is of dimension three since each one of its free indices range over the values 1, 2, 3 and hence it may be represented by a 3 × 3 matrix. However, in a 4D space this rank-2 tensor will be represented by a 4 × 4 matrix since its two free indices range over 1, 2, 3, 4. The rank of all terms in legitimate tensor expressions and equalities must be the same and hence: Aji − Bij and Aji = Cij (12) are legitimate but: Aji − Bi and jk Ajk i − Bi and j Ajim = Cim (14) jm Ajk i − Bi and j Ajim = Cin (15) Aji = C j (13) are illegitimate. Moreover, each term in valid tensor expressions and equalities must have the same set of free indices (e.g. i, j, k) and hence: are legitimate but: are illegitimate although they are all of the same rank. Also, a free index should keep its variance type in every term in valid tensor expressions and equations, i.e. it must be covariant in all terms or contravariant in all terms, and hence: mk Amk and Aji = Cij (16) iq − Biq are legitimate but: Aji − Bji and Ajin = Cijn (17) are illegitimate although they are all of the same rank and have the same set of free indices. We also note that the order of the tensor indices in legitimate tensor expressions and equalities should be the same and hence: Akm + Ckm and Djin = E jin (18) 1.2 General Background about Tensors 17 are legitimate but: i Aijk + Ckj and Dinm = Eimn (19) are illegitimate although they are all of the same rank, have the same set of free indices and have the same variance type. We remark that expressions and equalities like: Aij + Bji and iq qi Aiq k = B k − Ck (20) are common in the literature of tensor calculus and its applications which may suggest that the order of indices (or another one of the aforementioned features of the indicial structure) in tensor expressions and equalities is not important. However, expressions and equalities like these refer to the individual components of these tensors and hence they are of scalar, rather than tensor, nature. Hence the expression Aij + Bji means adding the value of the component Aij of tensor A to the value of the component Bji of tensor B and not adding iq qi the tensor A to the tensor B. Similarly, the equality Aiq k = Bk − Ck means subtracting the value of the component Ckiq of tensor C from the value of the component Bkqi of tensor B to obtain the value of the component Aiq k of tensor A. We may similarly write things i like Ai = B or Ai = Bj to mean the equality of the values of these components and not tensor equality. As we will see (refer to § 3.1.1), the indicial notation of tensors is made with reference to a set of basis vectors. For example, when we write Aijk as a tensor we mean Aijk Ei Ej Ek . This justifies all the above rules about the indicial structure (rank, set of free indices, variance type and order of indices) of tensor terms involved in tensor expressions and equalities because this structure is based on a set of basis vectors in a certain order. Therefore, an expression like Aijk + B ijk means Aijk Ei Ej Ek + B ijk Ei Ej Ek and an equation like Aijk = B ijk means Aijk Ei Ej Ek = B ijk Ei Ej Ek . However, as seen above it is common in the literature of tensor calculus that the tensor notation like Aijk is used to label the components and hence the above rules are not respected (e.g. Bij +Cji or ǫij = ǫij ) because these components are scalars in nature and hence these expressions and equalities do not refer to the vector basis. In this context, we should remark that an additional condition may be imposed on the indicial structure that is all the indices of a tensor should refer to a single vector basis set and hence a tensor cannot be indexed in reference to two different basis sets simultaneously.[3] While free indices should be named uniformly in all terms of tensor expressions and equalities, dummy indices can be named in each term independently and hence: b c Aaak + Bbk + Cck [3] and ja Dij = Eia + Fibjb (21) A tensor of different dimensionality (as in differential geometry) may not be compliant with this condition and hence we may add an extra condition that all the indices should refer to a single vector basis set for any particular space. Alternatively, a tensor like this can be seen as a tensor in reference to each individual type of indices separately but it is not a tensor simultaneously to both types of indices where the simultaneously condition takes care of this. 1.2 General Background about Tensors 18 are legitimate. The reason is that a dummy index represents a sum in its own term with no reach or presence into other terms. Despite the above restriction on the free indices, a free index in an expression or equality can be renamed uniformly and thoroughly using a different symbol, as long as this symbol is not already in use, assuming that both symbols vary over the same range, i.e. they have the same dimension. For example, we can change: jk Ajk i + Bi to jk Ajk p + Bp (22) j j − Dim Ajim = Cim to j j Ajpm = Cpm − Dpm (23) and change: as long as i and p have the same range and p is not already in use as an index for another purpose in that context. As indicated, the change should be thorough and hence all occurrences of the index i in that context, which may include other expressions and equalities, should be subject to that change. Regarding the dummy indices, they can be replaced by another symbol which is not present (as a free or dummy index) in their term as long as there is no confusion with a similar symbol in that context. Indexing is generally distributive over the terms of tensor expressions and equalities. For example, we have: [A + B]i = [A]i + [B]i (24) and [A = B]i ⇐⇒ [A]i = [B]i (25) Unlike scalars and tensor components, which are essentially scalars in a generic sense, operators cannot in general be freely reordered in tensor terms. Therefore, we have the following legitimate equalities: f h = hf and Ai B j = B j Ai (26) but we cannot equate ∂i Aj to Aj ∂i since in general we have: ∂i Aj 6= Aj ∂i (27) This should be obvious because ∂i Aj means that ∂i is operating on Aj but Aj ∂i means that ∂i is operating on something else and Aj just multiplies the result of this operation. As seen above, the order of the indices[4] of a given tensor is important and hence it should be observed and clarified, because two tensors with the same set of free indices and with the same indicial structure that satisfies the aforementioned rules but with different indicial order are not equal in general. For example, Aijk is not equal to Ajik unless A is symmetric with respect to the indices i and j (refer to § 3.1.5). Similarly, B mln is not equal to B lmn unless B is symmetric in its indices l and m. The confusion about the order of indices occurs specifically in the case of mixed type tensors such as Aijk which may not [4] The “order” here means “arrangement” and hence it should not be confused with the order of tensor as defined above in the same context as tensor rank. 1.3 Exercises and Revision 19 be clear since the order can be ijk or jik or jki. Spaces are usually used in this case to clarify the order. For example, the latter tensor is symbolized as Ai jk if the order of the indices is ijk, and as Aj i k if the order of the indices is jik while it is symbolized as Ajk i if the order of the indices is jki. Dots may also be used in such cases to indicate, more explicitly, the order of the indices and remove any ambiguity. For example, if the indices i, j, k of the tensor A, which is covariant in i and k and contravariant in j, are of that order, then A may be symbolized as Ai j. k where the dot between i and k indicates that j is in the middle. We note that in many places in this book (like many other books of tensor calculus) and mostly for the sake of convenience in typesetting, the order of the indices of mixed type tensors is not clarified by spacing or by inserting dots. This commonly occurs when the order of the indices is irrelevant in the given context (e.g. any order satisfies the intended purpose) or when the order is clear. Sometimes, the order of the indices may be indicated . jk implicitly by the alphabetical order of the selected indices, e.g. writing Ajk i to mean Ai . Finally, scalars, vectors and tensors may be defined on a single point of the space or on a set of separate points. They may also be defined over an extended continuous region (or regions) of the space. In the latter case we have scalar fields, vector fields and tensor fields, e.g. temperature field, velocity field and stress field respectively. Hence, a “field” is a function of coordinates which is defined over a given region of the space. As stated earlier, “tensor” may be used in a general sense to include scalar and vector and hence “tensor field” may include all the three types. 1.3 Exercises and Revision 1.1 Differentiate between the symbols used to label scalars, vectors and tensors of rank > 1. 1.2 What the comma and semicolon in Ajk,i and Ak;i mean? 1.3 State the summation convention and explain its conditions. To what type of indices this convention applies? 1.4 What is the number of components of a rank-3 tensor in a 4D space? 1.5 A symbol like Bijk may be used to represent tensor or its components. What is the difference between these two representations? Do the rules of indices apply to both representations or not? Justify your answer. 1.6 What is the meaning of the following symbols: ∇, ∂j , ∂kk , ∇2 , ∂φ , h,jk , Ai;n , ∂ n , ∇;k and Ci;km ? 1.7 What is the difference between symbolic notation and indicial notation? For what type of tensors these notations are used? What are the other names given to these types of notation? 1.8 “The characteristic property of tensors is that they satisfy the principle of invariance under certain coordinate transformations”. Does this mean that the components of tensors are constant? Why this principle is very important in physical sciences? 1.9 State and explain all the notations used to represent tensors of all ranks (rank-0, rank-1, rank-2, etc.). What are the advantages and disadvantages of using each one 1.3 Exercises and Revision 20 of these notations? 1.10 State the continuity condition that should be met if the equality: ∂i ∂j = ∂j ∂i is to be correct. 1.11 Explain the difference between free and bound tensor indices. Also, state the rules that govern each one of these types of index in tensor terms, expressions and equalities. 1.12 Explain the difference between the order and the rank of tensors and link this to the free and dummy indices. 1.13 What is the difference between covariant, contravariant and mixed type tensors? Give an example for each. 1.14 What is the meaning of “unit” and “zero” tensors? What is the characteristic feature of these tensors with regard to the value of their components? 1.15 What is the meaning of “orthonormal vector set” and “orthonormal coordinate system”? State any relevant mathematical condition. 1.16 What is the rule that governs the pair of dummy indices involved in summation regarding their variance type in general coordinate systems? Which type of coordinate system is exempt of this rule and why? 1.17 State all the rules that govern the indicial structure of tensors involved in tensor expressions and equalities (rank, set of free indices, variance type and order of indices). 1.18 How many equalities that the following equation contains assuming a 4D space: Bik = Cik ? Write all these equalities explicitly, i.e. B11 = C11 , B12 = C12 , etc. 1.19 Which of the following tensor expressions is legitimate and which is not, giving detailed explanation in each case? Aki − Bi , n − Bbb , Caa + Dm a + B, abj cdj + Fabk Scdk 1.20 Which of the following tensor equalities is legitimate and which is not, giving detailed explanation in each case? A.i n = B.ni , ab D = Scc + Nba , 3a + 2b = Jaa , k Bkm = Cm , Bj = 3c − Dj 1.21 Explain why the indicial structure (rank, set of free indices, variance type and order of indices) of tensors involved in tensor expressions and equalities are important referring in your explanation to the vector basis set to which the tensors are referred. Also explain why these rules are not observed in the expressions and equalities of tensor components. 1.22 Why free indices should be named uniformly in all terms of tensor expressions and equalities while dummy indices can be named in each term independently? 1.23 What are the rules that should be observed when replacing the symbol of a free index with another symbol? What about replacing the symbols of dummy indices? 1.24 Why in general we have: ∂i Aj 6= Aj ∂i ? What are the situations under which the following equality is valid: ∂i Aj = Aj ∂i ? 1.25 What is the difference between the order of a tensor and the order of its indices? 1.26 In which case Aijk is equal to Aikj ? What about Aijk and Aikj ? 1.27 What are the rank, order and dimension of the tensor Aijk in a 3D space? What about the scalar f and the tensor Aabm abjn from the same perspectives? 1.3 Exercises and Revision 21 1.28 What is the order of indices in Aj i k ? Insert a dot in this symbol to make the order more explicit. 1.29 Why the order of indices of mixed tensors may not be clarified by using spaces or inserting dots? 1.30 What is the meaning of “tensor field”? Is Ai a tensor field considering the spatial dependency of Ai and the meaning of “tensor”? Chapter 2 Spaces, Coordinate Systems and Transformations The focus of this chapter is coordinate systems, their types and transformations as well as some general properties of spaces which are needed for the development of the concepts and techniques of tensor calculus in the present and forthcoming chapters. The chapter also includes other sections which are intimately linked to these topics. 2.1 Spaces A Riemannian space is a manifold characterized by the existing of a symmetric rank-2 tensor called the metric tensor. The components of this tensor, which can be in covariant form gij or contravariant form g ij , as well as mixed form gji , are continuous variable functions of coordinates in general, that is: gij = gij (u1 , u2 , . . . , un ) g ij = g ij (u1 , u2 , . . . , un ) gji = gji (u1 , u2 , . . . , un ) (28) (29) (30) where the indexed u symbolizes general coordinates. This tensor facilitates, among other things, the generalization of the concept of length in general coordinate systems where the length of an infinitesimal element of arc, ds, is defined by: (ds)2 = gij dui duj (31) In the special case of a Euclidean space coordinated by an orthonormal Cartesian system, the metric becomes the identity tensor, that is: gij = δij g ij = δ ij gji = δji (32) More details about the metric tensor and its significance and roles will be given in § 4.5. The metric of a Riemannian space may be called the Riemannian metric. Similarly, the geometry of the space may be described as the Riemannian geometry. All spaces dealt with in the present book are Riemannian with well-defined metrics. As we will see, an nD manifold is Euclidean iff the Riemann-Christoffel curvature tensor vanishes identically (see § 7.2.1); otherwise the manifold is curved to which the general Riemannian geometry applies. In metric spaces, the physical quantities are independent of the form of description, being covariant or contravariant, as the metric tensor facilitates the transformation between the different forms; hence making the description objective. 22 2.1 Spaces 23 A manifold, such as a 2D surface or a 3D space, is called “flat” if it is possible to find a coordinate system for the manifold with a diagonal metric tensor whose all diagonal elements are ±1; the space is called “curved” otherwise. More formally, an nD space is described as flat space iff it is possible to find a coordinate system for which the length of an infinitesimal element of arc ds is given by: 2 1 2 2 2 n 2 (ds) = ζ1 (du ) + ζ2 (du ) + . . . + ζn (du ) = n X ζi (dui )2 (33) i=1 where the indexed ζ are ±1 while the indexed u are the coordinates of the space. For the space to be flat (i.e. globally not just locally), the condition given by Eq. 33 should apply all over the space and not just at certain points or regions. An example of flat space is the 3D Euclidean space which can be coordinated by an orthonormal Cartesian system whose metric tensor is diagonal with all the diagonal elements being +1. This also applies to plane surfaces which are 2D flat spaces that can be coordinated by 2D orthonormal Cartesian systems. Another example is the 4D Minkowski space-time manifold associated with the mechanics of Lorentz transformations whose metric is diagonal with elements of ±1 (see Eq. 241). When all the diagonal elements of the metric tensor of a flat space are +1, the space and the coordinate system may be described as homogeneous (see § 2.2.3). All 1D spaces are Euclidean and hence they cannot be curved intrinsically, so twisted curves are curved only when viewed externally from the embedding space which they reside in, e.g. the 2D space of a surface curve or the 3D space of a space curve. This is because any curve can be mapped isometrically to a straight line where both are naturally parameterized by arc length. An example of curved space is the 2D surface of a sphere or an ellipsoid since there is no possibility of coordinating these spaces with valid 2D coordinate systems that satisfy the above criterion. A curved space may have constant curvature all over the space, or have variable curvature and hence the curvature is position dependent. An example of a space of constant curvature is the surface of a sphere of radius R whose curvature (i.e. Riemannian curvature) is R12 at each point of the surface. Torus and ellipsoid are simple examples of 2D spaces with variable curvature. Schur theorem related to nD spaces (n > 2) of constant curvature states that: if the Riemann-Christoffel curvature tensor (see § 7.2.1) at each point of a space is a function of the coordinates only, then the curvature is constant all over the space. Schur theorem may also be stated as: the Riemannian curvature is constant over an isotropic region of an nD (n > 2) Riemannian space. A necessary and sufficient condition for an nD space to be intrinsically flat is that the Riemann-Christoffel curvature tensor of the space vanishes identically. Hence, cylinders are intrinsically flat, since their Riemann-Christoffel curvature tensor vanishes identically, although they are curved as seen extrinsically from the embedding 3D space. On the other hand, planes are intrinsically and extrinsically flat. In brief, a space is intrinsically flat iff the Riemann-Christoffel curvature tensor vanishes identically over the space, and it is extrinsically (as well as intrinsically) flat iff the curvature tensor vanishes identically over the whole space. This is because the Riemann-Christoffel curvature tensor characterizes 2.2 Coordinate Systems 24 the space curvature from an intrinsic perspective while the curvature tensor characterizes the space curvature from an extrinsic perspective. As indicated above, the geometry of curved spaces is usually described as the Riemannian geometry. One approach for investigating the Riemannian geometry of a curved manifold is to embed the manifold in a Euclidean space of higher dimensionality and inspect the properties of the manifold from this perspective. This approach is largely followed, for example, in the differential geometry of surfaces where the geometry of curved 2D spaces (twisted surfaces) is investigated by immersing the surfaces in a 3D Euclidean space and examining their properties as viewed from this external enveloping 3D space. Such an external view is necessary for examining the extrinsic geometry of the space but not its intrinsic geometry. A similar approach may also be followed in the investigation of surface and space curves. 2.2 Coordinate Systems In simple terms, a coordinate system is a mathematical device, essentially of geometric nature, used by an observer to identify the location of points and objects and describe events in generalized space which may include space-time. In tensor calculus, a coordinate system is needed to define non-scalar tensors in a specific form and identify their components in reference to the basis set of the system. An nD space requires a coordinate system with n mutually independent variable coordinates to be fully described so that any point in the space can be uniquely identified by the coordinate system. We note that the coordinates are generally real quantities although this may not apply in some cases (see § 2.2.3). As we will see in § 2.4, coordinate systems of 3D spaces are characterized by having coordinate curves and coordinate surfaces where the coordinate curves occur at the intersection of the coordinate surfaces.[5] The coordinate curves represent the curves along which exactly one coordinate varies while the other coordinates are held constant. Conversely, the coordinate surfaces represent the surfaces over which exactly one coordinate is held constant while the other coordinates vary. At any point P in a 3D space coordinated by a 3D coordinate system, we have 3 independent coordinate curves and 3 independent coordinate surfaces passing through P . The 3 coordinate curves uniquely identify the set of 3 mutually independent covariant basis vectors at P . Similarly, the 3 coordinate surfaces uniquely identify the set of 3 mutually independent contravariant basis vectors at P . Further details about this issue will follow in § 2.4. There are many types and categories of coordinate system; some of which will be briefly investigated in the following subsections. The most commonly used coordinate systems are: Cartesian, cylindrical and spherical. The most universal type of coordinate system is the general coordinate system which can include any type (rectilinear, curvilinear, orthogonal, etc.). A subset of the general coordinate system is the orthogonal coordinate system which is characterized by having mutually perpendicular coordinate curves, as well as mutually [5] In fact, these concepts can be generalized to nD spaces by generalizing the concepts of curves and surfaces. However, the main interest here and in the forthcoming sections is 3D spaces. 2.2.1 Rectilinear and Curvilinear Coordinate Systems 25 perpendicular coordinate surfaces, at each point in the region of space over which the system is defined and hence its basis vectors, whether covariant or contravariant, are mutually perpendicular. The coordinates of a system can have the same physical dimension or different physical dimensions. An example of the first is the Cartesian coordinate system, which is usually identified by (x, y, z), where all the coordinates have the dimension of length, while examples of the second include the cylindrical and spherical systems, which are usually identified by (ρ, φ, z) and (r, θ, φ) respectively, where some coordinates, like ρ and r, have the dimension of length while other coordinates, like φ and θ, are dimensionless. We also note that the physical dimensions of the components and basis vectors of the covariant and contravariant forms of a tensor are generally different. In the following subsections we outline a number of general types and categories of coordinate systems based on different classifying criteria. These categories are generally overlapping and may not be exhaustive in their domain. 2.2.1 Rectilinear and Curvilinear Coordinate Systems Rectilinear coordinate systems are characterized by the property that all their coordinate curves are straight lines and all their coordinate surfaces are planes, while curvilinear coordinate systems are characterized by the property that at least some of their coordinate curves are not straight lines and some of their coordinate surfaces are not planes (see § 2.4). Consequently, the basis vectors of rectilinear systems are constant while the basis vectors of curvilinear systems are variable in general since their direction or/and magnitude depend on the position in the space and hence they are coordinate dependent. Rectilinear coordinate systems can be rectangular (or orthogonal) when their coordinate curves, as well as their coordinate surfaces, are mutually orthogonal such as the well known rectangular Cartesian system. They can also be oblique when at least some of their coordinate curves and coordinate surfaces do not satisfy this condition. Figure 2 is a simple graphic illustration of rectangular and oblique rectilinear coordinate systems in a 3D space. Similarly, curvilinear coordinate systems can be orthogonal, when the vectors in their covariant or contravariant basis set are mutually orthogonal at each point in the space, and can be non-orthogonal when this condition is not met. Rectilinear coordinate systems may also be labeled as affine or linear coordinate systems although the terminology is not universal and hence these labels may be used differently. As stated above, curvilinear coordinate systems are characterized by the property that at least some of their coordinate curves are not straight lines and some of their coordinate surfaces are not planes (see Figures 3 and 10). This means that some (but not all) of the coordinate curves of curvilinear coordinate systems can be straight lines and some (but not all) of their coordinate surfaces can be planes. This is the case in the cylindrical and spherical coordinate systems as we will see next. Also, the coordinate curves of curvilinear coordinate systems may be regularly shaped curves such as circles and may be irregularly shaped and hence they are generalized twisted curves. Similarly, the coordinate surfaces of curvilinear systems may be regularly shaped surfaces such as spheres and may 2.2.2 Orthogonal Coordinate Systems 26 z z O y y x (a) O x (b) Figure 2: The two main types of rectilinear coordinate systems in 3D spaces: (a) rectangular and (b) oblique. be irregularly shaped. Prominent examples of curvilinear coordinate systems are the cylindrical and spherical systems of 3D spaces. All the coordinate curves and coordinate surfaces of these systems are regularly shaped. As we will see (refer to § 2.4), in the cylindrical coordinate systems the ρ, φ, z coordinate curves are straight lines, circles and straight lines respectively, while the ρ, φ, z coordinate surfaces are cylinders, semi-planes and planes respectively. Similarly, in the spherical coordinate systems the r, θ, φ coordinate curves are straight lines, semicircles and circles respectively, while the r, θ, φ coordinate surfaces are spheres, cones and semi-planes respectively. We note that an admissible coordinate transformation from a rectilinear system defines another rectilinear system if the transformation is linear, and defines a curvilinear system if the transformation is nonlinear. 2.2.2 Orthogonal Coordinate Systems The characteristic feature of orthogonal coordinate systems, whether rectilinear or curvilinear, is that their coordinate curves, as well as their coordinate surfaces, are mutually perpendicular at each point in their space. Hence, the vectors of their covariant basis set and the vectors of their contravariant basis set are mutually orthogonal. As a result, the corresponding covariant and contravariant basis vectors in orthogonal coordinate systems have the same direction and therefore if the vectors of these basis sets are normalized they will be identical, i.e. the normalized covariant and the normalized contravariant basis vector sets are the same. Prominent examples of orthogonal coordinate systems are rectangular Cartesian, cylindrical and spherical systems of 3D spaces (refer to Figure 4). A necessary and sufficient condition for a coordinate system to be orthogonal is that its metric tensor is diagonal. 2.2.3 Homogeneous Coordinate Systems 27 u3 E3 P u2 E2 E1 u1 Figure 3: General curvilinear coordinate system in a 3D space and its covariant basis vectors E1 , E2 and E3 (see § 2.6) as tangents to the shown coordinate curves at a particular point of the space P , where u1 , u2 and u3 represent general coordinates. This can be inferred from the definition of the components of the metric tensor as dot products of the basis vectors (see Eq. 48) since the dot product involving two different vectors (i.e. Ei · Ej or Ei · Ej with i 6= j) will vanish if the basis vectors, whether covariant or contravariant, are mutually perpendicular. 2.2.3 Homogeneous Coordinate Systems When all the diagonal elements of a diagonal metric tensor of a flat space are +1, the coordinate system is described as homogeneous. In this case the length of line element ds of Eq. 33 becomes: (ds)2 = dui dui (34) An example of homogeneous coordinate systems is the orthonormal Cartesian system of a 3D Euclidean space (Figure 4 a). A homogeneous coordinate system can be transformed to another homogeneous coordinate system only by linear transformations. Moreover, any coordinate system obtained from a homogeneous coordinate system by an orthogonal transformation (see § 2.3.3) is also homogeneous. As a consequence of the last statements, infinitely many homogeneous coordinate systems can be constructed in any flat space. A coordinate system of a flat space can always be homogenized by allowing the coordinates to be imaginary. This is done by redefining the coordinates as: p (no sum over i) (35) U i = ζ i ui where ζi = ±1. The new coordinates U i are real when ζi = 1 and imaginary when ζi = −1. 2.2.3 Homogeneous Coordinate Systems 28 z e3 O e1 e2 x y (a) Orthonormal Cartesian system and its orthonormal basis vectors e1 , e2 , e3 z z ez e P eρ φ O y φ ρ x (b) Cylindrical coordinate system and its orthonormal basis vectors eρ , eφ , ez z θ O φ er P eφ r eθ y x (c) Spherical coordinate system and its orthonormal basis vectors er , eθ , eφ Figure 4: The three prominent orthogonal coordinate systems in 3D spaces: (a) orthonormal Cartesian, (b) cylindrical and (c) spherical. 2.3 Transformations 29 Consequently, the length of line element ds will be given by: (ds)2 = dU i dU i (36) which is of the same form as Eq. 34. An example of a homogeneous coordinate system with some real and some imaginary coordinates is the coordinate system of a Minkowski 4D space-time of the mechanics of Lorentz transformations. We note that homogenization in the above sense is based on an extension of the concept of homogeneity and it is mainly based on the definition of the length of line element. 2.3 Transformations In general terms, a transformation from an nD space to another nD space is a correlation that maps a point from the first space (original) to a point in the second space (transformed) where each point in the original and transformed spaces is identified by n independent coordinates. To distinguish between the two sets of coordinates in the two spaces, the coordinates of the points in the transformed space may be notated with barred symbols like (ū1 , ū2 , . . . , ūn ), while the coordinates of the points in the original space are notated with unbarred similar symbols like (u1 , u2 , . . . , un ). Under certain conditions, which will be clarified later, such a transformation is unique and hence an inverse transformation from the transformed space to the original space is also defined. Mathematically, each one of the direct and inverse transformations can be regarded as a mathematical correlation expressed by a set of equations in which each coordinate in one space is considered as a function of the coordinates in the other space. Hence, the transformations between the two sets of coordinates in the two spaces can be expressed mathematically in a generic form by the following two sets of independent relations: ūi = ūi (u1 , u2 , . . . , un ) ui = ui (ū1 , ū2 , . . . , ūn ) (37) where i = 1, 2, . . . , n with n being the space dimension. The independence of the above relations is guaranteed iff the Jacobian of the transformation does not vanish at any point in the space (refer to the following paragraphs about the Jacobian). An alternative to the latter view of considering the transformation as a mapping between two different spaces is to view it as a correlation relating the same point in the same space but observed from two different coordinate systems which are subject to a similar transformation. The following will be largely based on the latter view although we usually adopt the one which is more convenient in the particular context. As far as the notation is concerned, there is no fundamental difference between the barred and unbarred systems and hence the notation can be interchanged. We also note that the transformations considered here, and in the present book in general, are between two spaces or coordinate systems of equal dimensions and hence we do not consider transformations between spaces or coordinate systems of different dimensions. Consequently, the Jacobian matrix of the transformation is always square and its determinant (i.e. the Jacobian) is defined. As indicated earlier, if the mapping from an original rectangular Cartesian system is linear, 2.3 Transformations 30 the coordinate system obtained from such a transformation is called affine or rectilinear. Coordinate systems which are not affine are described as curvilinear although the terminology may differ between the authors. The following n × n matrix of n2 partial derivatives of the unbarred coordinates with respect to the barred coordinates, where n is the space dimension, is called the “Jacobian matrix” of the transformation between the unbarred and barred systems:  ∂u1 ∂u1 1  · · · ∂∂uūn ∂ ū1 ∂ ū2  ∂u2 ∂u2 · · · ∂u2   1 ∂ ū2 ∂ ūn  (38) J =  ∂ ū. ..  . . . . .  . . .  . n ∂un ∂un · · · ∂u ∂ ū1 ∂ ū2 ∂ ūn while its determinant: J = det(J) (39) is called the “Jacobian” of the transformation where the indexed u and ū are the coordinates in the unbarred and barred coordinate systems of the nD space. The Jacobian array (whether matrix or determinant) contains all the possible n2 partial derivatives made of the different combinations of the u and ū indices where the pattern of the indices in this array is simple, that is the indices of u in the numerator provide the indices for the rows while the indices of ū in the denominator provide the indices for the columns. This labeling scheme may be interchanged which is equivalent to taking the transpose of the array. As it is well known, the Jacobian will not change by this transposition since the determinant of a matrix is the same as the determinant of its transpose, i.e. det(A) = det(AT ). We note that all coordinate transformations in the present book are continuous, single valued and invertible. We also note that “barred” and “unbarred” in the definition of Jacobian should be understood in a general sense not just as two labels since the Jacobian is not restricted to transformations between two systems of the same type but labeled as barred and unbarred. In fact the two coordinate systems can be fundamentally different in nature such as orthonormal Cartesian and general curvilinear. The Jacobian matrix and determinant represent any transformation by the above partial derivative array system between two coordinate systems defined by two different sets of coordinate variables not necessarily as barred and unbarred. The objective of defining the Jacobian as between unbarred and barred systems is simplicity and generality. The transformation from the unbarred coordinate system to the barred coordinate system is bijective[6] iff J 6= 0 at any point in the transformed region of the space. In this case, the inverse transformation from the barred to the unbarred system is also defined and bijective and is represented by the inverse of the Jacobian matrix, that is: J̄ = J−1 (40) Consequently, the Jacobian of the inverse transformation, being the determinant of the inverse Jacobian matrix, is the reciprocal of the Jacobian of the original transformation, [6] Bijective means the mapping is injective (one-to-one) and surjective (onto). 2.3 Transformations 31 that is: 1 (41) J¯ = J As we remarked, there is no fundamental notational difference between the barred and unbarred systems and hence the labeling is rather arbitrary and can be interchanged. Therefore, the Jacobian may be notated as unbarred over barred or the other way around. The essence is that the Jacobian usually represents the transformation from an original system to another system while its inverse represents the opposite transformation although even this is not generally respected in the literature of mathematics and hence “Jacobian” my be used to label the opposite transformation. Yes, in a specific context when one of these is labeled as the Jacobian, the other one should be labeled as the inverse Jacobian to distinguish between the two opposite Jacobians and their corresponding transformations. However, there may also be practical aspects for choosing which is the Jacobian and which is the inverse since it is easier sometimes to compute one of these than the other and hence we start by computing the easier as the Jacobian (whether from original to transformed or the other way) followed by obtaining the reciprocal as the inverse Jacobian. Anyway, in this book we generally use “Jacobian” flexibly where the context determines the nature of the transformation.[7] An admissible (or permissible or allowed) coordinate transformation may be defined generically as a mapping represented by a sufficiently differentiable set of equations plus being invertible by having a non-vanishing Jacobian (J 6= 0). More technically, a coordinate transformation is commonly described as admissible iff the transformation is bijective with non-vanishing Jacobian and the transformation function is of class C 2 .[8] We note that the C n continuity condition means that the function and all its first n partial derivatives do exist and are continuous in their domain. Also, some authors may impose a weaker continuity condition of being of class C 1 . An object that does not change by admissible coordinate transformations is described as “invariant” such as a true scalar (see § 3.1.2) which is characterized by its sign and magnitude and a true vector which is characterized by its magnitude and direction in space. Similarly, an invariant property of an object or a manifold is a property that is independent of admissible coordinate transformations such as being form invariant which characterizes tensors or being flat which characterize spaces under certain types of transformation. It should be noted that an invariant object or property may be invariant with respect to certain types of transformation but not with respect to other types of transformation and hence the term may be used generically where the context is taken into consideration for [7] We may get rid of all these complications by using the term “Jacobian” to represent a system of partial derivatives whose objective is to transform from one coordinate system to another (regardless of anything else like system notation or being original and secondary) and hence both the “Jacobian” and its inverse are Jacobians in this general sense; the two will be distinguished from each other by stating from which system to which system the Jacobian transforms. Hence, we can say legitimately and with no ambiguity: the Jacobian from A to B and the Jacobian from B to A. These may be labeled as J (A → B) and J (B → A) with similar notation for the Jacobian matrix. This, we believe, can resolve all these issues and avoid confusion. [8] The meaning of “admissible coordinate transformation” may vary depending on the context. 2.3.1 Proper and Improper Transformations 32 sensible interpretation. A product or composition of space or coordinate transformations is a succession of transformations where the output of one transformation is taken as the input to the next transformation. For example, a series of m transformations labeled as Ti where i = 1, 2, · · · , m may be applied sequentially onto a mathematical object O. This operation can be expressed mathematically by the following composite transformation Tc : Tc (O) = Tm Tm−1 · · · T2 T1 (O) (42) where the output of T1 in this notation is taken as the input to T2 and so forth until the output of Tm−1 is fed as an input to Tm in the end to produce the final output of Tc . In such cases, the Jacobian of the product is the product of the Jacobians of the individual transformations of which the product is made, that is: Jc = Jm Jm−1 · · · J2 J1 (43) where Ji in this notation is the Jacobian of the Ti transformation and Jc is the Jacobian of Tc . The collection of all admissible coordinate transformations with non-vanishing Jacobian form a group. This means that they satisfy the properties of closure, associativity, identity and inverse. Hence, any convenient coordinate system can be chosen as the point of entry since other systems can be reached, if needed, through the set of admissible transformations. This is one of the cornerstones of building invariant physical theories which are independent of the subjective choice of coordinate systems and reference frames. We remark that transformation of coordinates is not a commutative operation and hence the result of two successive transformations may depend on the order of these transformations. This is demonstrated in Figure 5 where the composition of two rotations results in different outcomes depending on the order of the rotations. As there are essentially two different types of basis vectors, namely tangent vectors of covariant nature and gradient vectors of contravariant nature (see § 2.6 and 3.1.1), there are two main types of non-scalar tensors: contravariant tensors and covariant tensors which are based on the type of the employed basis vectors of the given coordinate system. Tensors of mixed type employ in their definition mixed basis vectors of the opposite type to the corresponding indices of their components. As we will see, the transformation between these different types is facilitated by the metric tensor of the given coordinate system (refer to § 4.5). In the following subsections, we briefly describe a number of types and categories of coordinate transformations. 2.3.1 Proper and Improper Transformations Coordinate transformations are described as “proper” when they preserve the handedness (right- or left-handed) of the coordinate system and “improper” when they reverse the handedness. Improper transformations involve an odd number of coordinate axes inversions in the origin of coordinates. Inversion of axes may be called improper rotation while 2.3.2 Active and Passive Transformations 33 y z y z y x x z z y z x x z y x y x Figure 5: The composite transformations Ry Rx (top) and Rx Ry (bottom) where Rx and Ry represent clockwise rotation of π2 around the positive x and y axes respectively. ordinary rotation is described as proper rotation. Figure 6 illustrates proper and improper coordinate transformations of a rectangular Cartesian coordinate system in a 3D space. 2.3.2 Active and Passive Transformations Transformations can be active, when they change the state of the observed object such as rotating the object in the space, or passive when they are based on keeping the state of the object and changing the state of the coordinate system which the object is observed from. In brief, the subject of an active transformation is the object while the subject of a passive transformation is the coordinate system. 2.3.3 Orthogonal Transformations An orthogonal coordinate transformation consists of a combination of translation, rotation and reflection of axes. The Jacobian of orthogonal transformations is unity, that is J = ±1.[9] The orthogonal transformation is described as positive iff J = +1 and neg[9] This condition should apply even when the transformation includes a translation since the added constants that represent the translation in the transformation equations will vanish in the Jacobian matrix. However, there seems to be a different convention that excludes translation to be part of orthogonal transformations. There also seems to be another convention which restricts orthogonal transformations to translation and rotation. 2.3.4 Linear and Nonlinear Transformations 34 x3 X3 x3 X3 O X1 O X2 x1 X1 Proper x2 X2 x1 Improper x2 Figure 6: Proper and improper transformations of a rectangular Cartesian coordinate system in a 3D space where the former is achieved by a rotation of the coordinate system while the latter is achieved by a rotation followed by a reflection of the first axis in the origin of coordinates. The transformed systems are shown as dashed and labeled with upper case letters while the original system is shown as solid and labeled with lower case letters. ative iff J = −1. Positive orthogonal transformations consist solely of translation and rotation (possibly trivial ones as in the case of the identity transformation) while negative orthogonal transformations include reflection, by applying an odd number of axes reversal, as well. Positive transformations can be decomposed into an infinite number of continuously varying infinitesimal positive transformations each one of which imitates an identity transformation. Such a decomposition is not possible in the case of negative orthogonal transformations because the shift from the identity transformation to reflection is impossible by a continuous process. 2.3.4 Linear and Nonlinear Transformations The characteristic property of linear transformations is that they maintain scalar multiplication and algebraic addition, while nonlinear transformations do not. Hence, if T is a linear transformation then we have: T (aA ± bB) = a T (A) ± b T (B) (44) where A and B are mathematical objects to be transformed by T and a and b are scalars. As indicated earlier, an admissible coordinate transformation from a rectilinear system defines another rectilinear system if the transformation is linear, and defines a curvilinear system if the transformation is nonlinear. 2.4 Coordinate Curves and Coordinate Surfaces 2.4 35 Coordinate Curves and Coordinate Surfaces As seen earlier, coordinate systems of 3D spaces are characterized by having coordinate curves and coordinate surfaces where the coordinate curves represent the curves of mutual intersection of the coordinate surfaces.[10] These coordinate curves and coordinate surfaces play a crucial role in the formulation and development of the mathematical structures of the coordinated space. The coordinate curves represent the curves along which exactly one coordinate varies while the other coordinates are held constant. Conversely, the coordinate surfaces represent the surfaces over which all coordinates vary except one which is held constant. In brief, the ith coordinate curve is the curve along which only the ith coordinate varies while the ith coordinate surface is the surface over which only the ith coordinate is constant. For example, in a 3D Cartesian system identified by the coordinates (x, y, z) the curve r(x) = (x, c2 , c3 ), where c2 and c3 are real constants, is an x coordinate curve since x varies while y and z are held constant, and the surface r(x, z) = (x, c2 , z) is a y coordinate surface since x and z vary while y is held constant. Similarly, in a cylindrical coordinate system identified by the coordinates (ρ, φ, z) the curve r(φ) = (c1 , φ, c3 ), where c1 is a real constant, is a φ coordinate curve since ρ and z are held constant while φ varies. Likewise, in a spherical coordinate system identified by the coordinates (r, θ, φ) the surface r(θ, φ) = (c1 , θ, φ) is an r coordinate surface since r is held constant while θ and φ vary. As stated before, coordinate curves represent the curves of mutual intersection of coordinate surfaces. This should be obvious since along the intersection curve of two coordinate surfaces, where on each one of these surfaces one coordinate is held constant, two coordinates will be constant and hence only the third coordinate can vary. Hence, in a 3D Cartesian coordinate system the x coordinate curves occur at the intersection of the y and z coordinate surfaces, the y coordinate curves occur at the intersection of the x and z coordinate surfaces, and the z coordinate curves occur at the intersection of the x and y coordinate surfaces (refer to Figure 7). Similarly, in a cylindrical coordinate system the ρ, φ and z coordinate curves occur at the intersection of the φ, z, the ρ, z and the ρ, φ coordinate surfaces respectively (refer to Figure 8). Likewise, in a spherical coordinate system the r, θ and φ coordinate curves occur at the intersection of the θ, φ, the r, φ and the r, θ coordinate surfaces respectively (refer to Figure 9). In this context, we note that the three types of coordinate surface of a Cartesian system, or in fact any rectilinear coordinate system (see § 2.2.1), are planes and hence the three types of coordinate curves are straight lines (refer to Figure 7). As for cylindrical coordinate systems, the ρ coordinate surfaces are cylinders (and this maybe the reason behind labeling them as “cylindrical”), the φ coordinate surfaces are semi-planes and the z coordinate surfaces are planes. Hence, the ρ, φ and z coordinate curves are straight lines, circles, and straight lines respectively (refer to Figure 8). Regarding spherical coordinate systems, the r coordinate surfaces are spheres, the θ coordinate surfaces are cones and the φ coordinate surfaces are semi-planes. Hence, the r, θ and φ coordinate curves are [10] Our focus here is 3D spaces although these concepts can be generalized to nD spaces as indicated earlier. 36 yC S 2.5 Scale Factors z z CC y CC z CS xC C x CS y x Figure 7: Coordinate curves (CC) and coordinate surfaces (CS) in a 3D Cartesian coordinate system. straight lines, semi-circles, and circles respectively (refer to Figure 9). The coordinate surfaces of general curvilinear coordinate systems can vary in general and hence they are not represented by particular geometric shapes such as planes and spheres. As a result, the coordinate curves of these systems are general space curves which may not have regular geometric shapes such as straight line or circle. As stated earlier, orthogonal coordinate systems are characterized by having coordinate surfaces which are mutually orthogonal at each point of the space, and consequently their coordinate curves are also mutually orthogonal at each point of the space. Finally, we remark that the transformation relations of Eq. 37 are used in defining the set of coordinate surfaces and coordinate curves. 2.5 Scale Factors Scale factors of a coordinate system are those factors which are required to multiply the coordinate differentials to obtain the distances traversed during a change in the coordinate of that magnitude. The scale factors are symbolized with h1 , h2 , . . . , hn where this notation is usually used in orthogonal coordinate systems. For example, in the plane polar coordinate system represented by the coordinates (ρ, φ), the scale factor of the second coordinate φ is ρ because ρ is the factor used to multiply the differential of the polar angle dφ to obtain the distance L traversed by a change of magnitude dφ in the polar angle, 2.5 Scale Factors 37 z ρ CC z CS z O φ CS φ y x z (a) ρ coordinate curvex with φ and z coordinate surfaces z z z CS z ρ ρ y φ CC O ρ CS y x z x (b) φ coordinate curve with ρ zand z coordinate surfaces z φ CS ρ CS y O ρ x φ ρ ρ y y x z CC x (c) z coordinate curve with ρ and φ coordinate surfaces Figure 8: Coordinate curves (CC) and coordinate surfaces(CS) in cylindrical coordinate systems. 2.5 Scale Factors 38 z θ CS r CC θ O φ φ CS y x (a) r coordinate curve with zθ and φ coordinate surfaces z r CS θ CC O r r φ x x y y φ CS z (b) θ coordinate curve withzr and φ coordinate surfaces z CS r θ CS r r θ y y y φ CC O x x x r CS (c) φ coordinate curve with r and θ coordinate surfaces Figure 9: Coordinate curves (CC) and coordinate surfaces (CS) in spherical coordinate systems. 2.6 Basis Vectors and Their Relation to the Metric and Jacobian 39 Table 1: The scale factors (h1 , h2 , h3 ) for the three most commonly used orthogonal coordinate systems in 3D spaces: orthonormal Cartesian, cylindrical and spherical. The squares of these entries and the reciprocals of these squares give the diagonal elements of the covariant and contravariant metric tensors, gij and g ij , respectively of these systems (see Eqs. 238-240). h1 h2 h3 Cartesian (x, y, z) 1 1 1 Cylindrical (ρ, φ, z) 1 ρ 1 Spherical (r, θ, φ) 1 r r sin θ that is: L = ρ dφ. The scale factors are also used for other purposes such as normalizing the basis vectors and defining the components in the physical representation of vectors and tensors (refer to § 2.6 and 3.3). They are also used in the analytical expressions for length, area and volume in orthogonal coordinate systems, as described in § 4.6.8, 4.6.9 and 4.6.10. The scale factors for the Cartesian, cylindrical and spherical coordinate systems of 3D spaces are given in Table 1. 2.6 Basis Vectors and Their Relation to the Metric and Jacobian The set of basis vectors in a given manifold plays a pivotal role in the theoretical construction of the geometry of the manifold, where these vectors are used in the definition and construction of essential concepts and objects such as the metric tensor of the space. The basis sets are defined at each regular point in the space and hence the vectors in the basis sets vary in general from one point to another, i.e. they are coordinate dependent. The vectors providing the basis set for a coordinate system, which are not necessarily of unit length or mutually orthogonal, can be of covariant type or contravariant type. The covariant basis vectors are defined as the tangent vectors to the coordinate curves, while the contravariant basis vectors are defined as the gradient of the space coordinates and hence they are perpendicular to the coordinate surfaces (refer to Figure 10). Formally, the covariant and contravariant basis vectors are defined respectively by: Ei = ∂r ∂ui Ei = ∇ui (45) where r is the position vector in Cartesian coordinates (x1 , . . . , xn ), ui represents general coordinates, n is the space dimension and i = 1, · · · , n. As indicated above, the covariant and contravariant basis sets, Ei and Ei , in general coordinate systems are functions of coordinates, that is:   (46) E i = Ei u1 , . . . , u n E i = Ei u1 , . . . , u n Hence, the definitions of Eq. 45 apply to each individual point in the space where the coordinate curves and coordinate surfaces belong to that particular point. For example, 2.6 Basis Vectors and Their Relation to the Metric and Jacobian 40 at any particular point P in a 3D space with a valid coordinate system we have three mutually independent coordinate curves and three mutually independent coordinate surfaces and hence we should have three mutually independent covariant basis vectors and three mutually independent contravariant basis vectors at P . u3 CC u1 CS E3 E3 2 u CS P E2 E1 u2 CC E2 E1 u1 CC u3 CS Figure 10: The covariant and contravariant basis vectors of a general curvilinear coordinate system and the associated coordinate curves (CC) and coordinate surfaces (CS) at a given point P in a 3D space. Like other vectors, the covariant and contravariant basis vectors of a given coordinate system are related to each other, through the metric tensor of the system, by the following relations: Ei = gij Ej Ei = g ij Ej (47) where the metric tensor in its covariant form gij and contravariant form g ij works as an index shifting operator to lower and raise the indices and hence change the variance type of the basis vectors. Thus, the transformation between the covariant basis set and the contravariant basis set of a particular coordinate system is facilitated by the metric tensor of that system. The basis vectors in their covariant and contravariant forms are related to the components of the metric tensor in its covariant and contravariant forms by the following relations: Ei · Ej = gij Ei · Ej = g ij (48) 2.6 Basis Vectors and Their Relation to the Metric and Jacobian 41 The covariant and contravariant basis vectors are reciprocal basis systems, and hence we have: Ei · Ej = δij Ei · Ej = δ ij (49) where the indexed δ is the Kronecker delta tensor (refer to § 4.1). As we will see in § 4.5, the Kronecker delta tensor represents the metric tensor in its mixed form. In a 3D space with a right handed coordinate system the two sets of basis vectors are linked by the following relations: E1 = E2 × E3 , E1 · (E2 × E3 ) E2 = E3 × E1 , E1 · (E2 × E3 ) E3 = E1 × E2 E1 · (E2 × E3 ) (50) E3 × E1 E1 × E2 E2 × E3 , E2 = , E3 = (51) E1 · (E2 × E3 ) E1 · (E2 × E3 ) E1 · (E2 × E3 ) The pattern in these relations is very simple, that is the covariant vectors are represented by contravariant vectors and vice versa; moreover, the indices in the numerators on both sides represent even permutations of 1, 2, 3. The relations in Eqs. 50 and 51 may be expressed in a more compact form as follows: E1 = Ei = Ej × Ek Ei · (Ej × Ek ) Ei = Ej × Ek Ei · (Ej × Ek ) (52) where i, j, k take respectively the values 1, 2, 3 and the other two cyclic permutations (i.e. 2, 3, 1 and 3, 1, 2). It is worth noting that the magnitude of the scalar triple product Ei ·(Ej × Ek ) represents the volume of the parallelepiped formedby the vectors Ei , Ej and Ek while the magnitude of the scalar triple product Ei · Ej × Ek represents its reciprocal. We also note that the scalar triple product is invariant to a cyclic permutation of the symbols of the three vectors involved, i.e. Ei · (Ej × Ek ) = Ek · (Ei × Ej ) = Ej · (Ek × Ei ); in fact its magnitude is invariant even to non-cyclic permutations. Another important note is that the magnitudes of the basis vectors in orthogonal coordinate systems are related to the scale factors of the coordinate system by: |Ei | = hi Ei = 1 hi (53) where hi is the scale factor for the ith coordinate (see § 2.5). Following an admissible coordinate transformation between unbarred and barred general coordinate systems, the basis vectors in these systems are related by the following transformation rules: ∂ ūj Ēj ∂ui ∂ui j Ei = Ē ∂ ūj Ei = ∂uj Ej ∂ ūi ∂ ūi j E Ēi = ∂uj Ēi = (54) (55) where the indexed u and ū represent the coordinates in the unbarred and barred systems respectively. The transformation rules for the components can be easily deduced from 2.6 Basis Vectors and Their Relation to the Metric and Jacobian 42 the above rules. For example, for a vector A which can be represented covariantly and contravariantly in the unbarred and barred systems as: A = Ei Ai = Ēi Āi (56) A = Ei Ai = Ēi Āi (57) the transformation equations of its components between the two systems are given respectively by: ∂ ūj Āj ∂ui ∂ui j Ā Ai = ∂ ūj ∂uj Aj ∂ ūi ∂ ūi j Āi = A ∂uj Ai = (58) Āi = (59) These transformation rules can be easily extended to higher rank tensors of different variance types (see § 3.1.1). For a 3D manifold with a right handed coordinate system, we have: E1 · (E2 × E3 ) = √  1 E1 · E2 × E3 = √ g g (60) where g is the determinant of the covariant metric tensor, i.e. (61) g = det (gij ) = |gij | Because Ei · Ej = gij (Eq. 48) we have:[11] JT J = [gij ] (62) where J is the Jacobian matrix that transforms between Cartesian and general coordinates, the superscript T represents matrix transposition, [gij ] is the matrix representing the covariant metric tensor and the product on the left is a matrix product as defined in linear algebra which is equivalent to a dot product in tensor algebra. On taking the determinant of both sides of Eq. 62, the relation between the Jacobian and the determinant of the metric tensor is obtained, that is: g = J2 (63) where J is the Jacobian of the transformation. As it should be known, the determinant of a product is equal to the product of the determinants, that is: det(JT J) = det(JT )det(J) [11] (64) For example, for the transformation between Cartesian and general coordinate systems in 3D spaces we have:     1 2 3 1 1 1 T  J = ∂x ∂u11 ∂x ∂u12 ∂x ∂u3 ∂x ∂u21 ∂x ∂u22 ∂x ∂u3 ∂x ∂u31 ∂x ∂u32 ∂x ∂u3    J= ∂x ∂u21 ∂x ∂u31 ∂x ∂u1 ∂x ∂u22 ∂x ∂u32 ∂x ∂u2 ∂x ∂u23 ∂x ∂u33 ∂x ∂u3   The entries of the product of these matrices then correspond to the entries of the metric tensor as given by Eq. 216. 2.6 Basis Vectors and Their Relation to the Metric and Jacobian 43 Moreover, the determinant of a matrix is equal to the determinant of its transpose, that is: det(JT ) = det(J) = J (65) As explained earlier, in orthogonal coordinate systems the covariant and contravariant basis vectors, Ei and Ei , at any point of the space are in the same direction, and hence the normalization of each one of these basis sets, by dividing each basis vector by its magnitude, produces identical orthonormal basis sets. Consequently, there is no difference between the covariant and contravariant components of tensors with respect to such contravariant and covariant orthonormal basis sets. This, however, is not true in general coordinate systems where each normalized basis set is different in general from the other. When the covariant basis vectors Ei are mutually orthogonal at each point of the space, the following consequences will follow: 1. The contravariant basis vectors Ei are also mutually orthogonal because the corresponding vectors of each basis set are in the same direction due to the fact that the tangent vector to the ith coordinate curve and the gradient vector of the ith coordinate surface at a given point in the space have the same direction. This may also be established more formally by Eq. 49 since the reciprocal of an orthogonal system of vectors should also be orthogonal. Eqs. 50 and 51 may also be used in this argument since the cross products in the numerators are orthogonal to the multiplicands. 2. The covariant and contravariant metric tensors, gij and g ij , are diagonal with nonvanishing diagonal elements, that is:[12] gij = 0 g ij = 0 gii 6= 0 g ii 6= 0 (i 6= j) (no sum on i) (66) (67) This can be concluded from Eq. 48 since the dot product is zero when the indices are different due to orthogonality. Also, the dot product should be non-zero when the indices are identical because the basis vectors cannot vanish at the regular points of the space since the tangent to the coordinate curve and the gradient to the coordinate surface do exist and they cannot be zero. Moreover, the metric tensor (as we will see in § 4.5) is invertible and hence it does not vanish at any point in the space. This guarantees that none of the diagonal elements can be zero when the metric tensor is diagonal since the determinant, which is formed by the product of its diagonal elements, should not vanish. 3. The diagonal elements of the covariant and contravariant metric tensors are reciprocals, that is: 1 (no summation) (68) g ii = gii This is a result from the previous point (i.e. the metric tensor is diagonal with nonvanishing diagonal elements) since the covariant and contravariant metric tensors [12] This also applies to the mixed type metric tensor since it is the identity tensor according to Eq. 49 although this is not restricted to orthogonal coordinate systems. 2.7 Relationship between Space, Coordinates and Metric 44 are inverses of each other. As it is well known from linear algebra, the inverse of a diagonal matrix with non-vanishing diagonal entries is a diagonal matrix obtained by taking the reciprocals of its diagonal elements. 4. The magnitudes of the contravariant and covariant basis vectors are reciprocals, that is: 1 1 Ei = = (69) |Ei | hi This result is based on the definition of vector modulus plus Eqs. 48, 68 and 53, that is: √ Ei = Ei · Ei (Eq. 263) (70) p (Eq. 48) = g ii 1 =√ (Eq. 68) gii 1 (Eq. 48) =√ Ei · Ei 1 = (Eq. 264) |Ei | 1 = (Eq. 53) hi with no sum over i. The consequences that we stated in the above bullet points also follow when the contravariant basis vectors Ei are mutually orthogonal with minor amendments required in the phrasing of the first point. 2.7 Relationship between Space, Coordinates and Metric In this section, we try to shed some light on the relationship between space, coordinate system and metric tensor where in this discussion we need the concepts of basis vectors and coordinate curves and coordinate surfaces (see § 2.4 and 2.6).[13] We start from the very abstract concept of space (say nD space) which may be visualized as a sort of vacuum in which mathematical objects can exist. The number n represents the dimensionality of the space which is a quantitative measure of the complexity of the space structure where an nD space requires exactly n mutually independent variables to be fully and uniquely described. So, any point in this nD space can be uniquely located by assigning an ordered set of n mutually independent variables which we call “coordinates” such as u1 , · · · , un . These coordinates vary continuously throughout the space where each coordinate ranges over a given real interval which may be finite or infinite. All the points of the space are then uniquely identified by varying these variables over their predefined ranges. The way [13] The revelation in this section is highly informal and non-rigorous with a pedagogical objective of providing a qualitative appreciation of the relation between space, coordinate system and metric tensor which may be useful for some novice readers. 2.7 Relationship between Space, Coordinates and Metric 45 by which the points of space are identified by the ordered sets of n coordinates may be labeled generically as a “coordinate system” for the space. The condition of mutual independence of the coordinate variables is important to make the distinction between the points unique and universal and to avoid any ambiguity. Mutual independence requires that an nD space is identified by a coordinate system with exactly n variable coordinates. As well as the condition of mutual independence, there is another important condition for the coordinate system to be acceptable that is it should be universal in its ability to identify all the points of the space unambiguously, and hence the coordinate system should be thorough in its functionality with possible exclusion of a finite number of isolated points in the space. The “coordinate system” may then be defined more rigorously as a bijective correspondence between the set of n-tuple numbers and the points of the space. For the coordinate system to be of any use, the way by which the coordinates are assigned to the individual points should be clearly defined. So, each particular way of defining the aforementioned bijective correspondence represents a specific coordinate system associated with the space. In fact, there are infinitely many ways of choosing the coordinate variables and how they vary in the space, and hence there are infinitely many coordinate systems that can be chosen to fulfill the above objectives. However, the conditions of mutual independence of coordinates and bijectivity of mapping impose certain restrictions on the selection of coordinates and how they can vary to achieve the intended functionalities. As a result of the above conditions and restrictions on the coordinate variables and how they vary to ensure their mutual independence and universal validity at each point in the space, there are only certain allowed ways for the choice of the coordinate curves and coordinate surfaces. So, on excluding certain types of coordinate systems which do not meet these requirements, and hence excluding certain types of how the coordinate curves and surfaces are defined at the individual points of the space, we are left with only certain acceptable possibilities for identifying how the coordinates vary and hence how the coordinate curves and surfaces are identified. Now, since the covariant basis vectors are defined as the tangents to the coordinate curves and the contravariant basis vectors are defined as the gradients to the coordinate surfaces, then to each one of those acceptable possibilities there is only one possibility for the identification of the covariant and contravariant sets of basis vectors. As soon as the two sets of basis vectors of the selected coordinate system are identified unambiguously, the metric tensor in its covariant and contravariant forms is uniquely identified throughout the space through its relation to the basis vectors, i.e. Eq. 48, and hence the vagueness about the nature of the space, its geometry and how it should be described is removed thanks to the existence of a well defined metric tensor. In brief, we start by defining a coordinate system which leads to the definition of coordinate curves and coordinate surfaces (and hence covariant and contravariant basis vector sets) and this in its turn will lead to the definition of the metric tensor. In fact, this approach may be described as procedural since it is based on the practical procedure of defining these concepts. A more fundamental approach may start from the metric of the space as an essential property of the space that characterizes its nature from which the 2.8 Exercises and Revision 46 basis vectors and the coordinate system are to be derived. It should be remarked that the coordinate system and the space metric are independent entities despite the close link between the two. The metric can be regarded as a built-in attribute that characterizes the space but it can be demonstrated in different forms depending on the employed coordinate system. In this context, it is important to distinguish between the intrinsic and extrinsic properties of an nD space where the intrinsic properties refer to those properties of the space which are seen from inside the space by an inhabitant of the space who can perceive only the n dimensions of the space, while the extrinsic properties refer to those properties of the space which are seen from outside the space by an inhabitant of an mD space (m > n) that encloses the nD space where this mD inhabitant can perceive the n dimensions of the space plus the extra (m − n) dimensions of the mD space. While the intrinsic properties of the nD space are described and quantified by the metric of the nD space, the extrinsic properties of the space are described and quantified by the metric of the embedding mD space since the nD space is a subspace of the mD space and hence it is subject to the mD metric. An example of nD and mD spaces is a 2D surface, such as a plane or a sphere, which is embedded in a 3D Euclidean space. 2.8 Exercises and Revision 2.1 Give brief definitions to the following terms: Riemannian space, coordinate system and metric tensor. 2.2 Discuss the main functions of the metric tensor in a given space. How many types the metric tensor can have? 2.3 What is the meaning of “flat” and “curved” space? Give mathematical conditions for the space to be flat in terms of the length of an infinitesimal element of arc and in terms of the metric tensor. Why these conditions should be global for the space to be flat? 2.4 Give common examples of flat and curved spaces of different dimensions justifying in each case why the space is flat or curved. 2.5 Explain why all 1D spaces are Euclidean. 2.6 Give examples of spaces with constant curvature and spaces with variable curvature. 2.7 State Schur theorem outlining its significance. 2.8 What is the condition for a space to be intrinsically flat and extrinsically flat? 2.9 What is the common method of investigating the Riemannian geometry of a curved manifold? 2.10 Give brief definitions to coordinate curves and coordinate surfaces outlining their relations to the basis vector sets. How many independent coordinate curves and coordinate surfaces we have at each point of a 3D space with a valid coordinate system? 2.11 Why a coordinate system is needed in tensor formulations? 2.12 List the main types of coordinate system outlining their relations to each other. 2.13 “The coordinates of a system can have the same physical dimension or different phys- 2.8 Exercises and Revision 2.14 2.15 2.16 2.17 2.18 2.19 2.20 2.21 2.22 2.23 2.24 2.25 2.26 2.27 2.28 2.29 2.30 2.31 2.32 2.33 2.34 2.35 2.36 2.37 2.38 47 ical dimensions”. Give an example for each. Prove that spherical coordinate systems are orthogonal. What is the difference between rectilinear and curvilinear coordinate systems? Give examples of common curvilinear coordinate systems explaining why they are curvilinear. Give an example of a commonly used curvilinear coordinate system with some of its coordinate curves being straight lines. Define briefly the terms “orthogonal” and “homogeneous” coordinate system. Give examples of rectilinear and curvilinear orthogonal coordinate systems. What is the condition of a coordinate system to be orthogonal in terms of the form of its metric tensor? Explain why this is so. What is the mathematical condition for a coordinate system to be homogeneous? How can we homogenize a non-homogeneous coordinate system of a flat space? Give examples of homogeneous and non-homogeneous coordinate systems. Give an example of a non-homogeneous coordinate system that can be homogenized. Describe briefly the transformation of spaces and coordinate systems stating relevant mathematical relations. What “injective transformation” means? Is it necessary that such a transformation has an inverse? Write the Jacobian matrix J of a transformation between two nD spaces whose coordinates are labeled as ui and ūi where i = 1, · · · , n. State the pattern of the row and column indices of the Jacobian matrix in relation to the indices of the coordinates of the two spaces. What is the difference between the Jacobian matrix and the Jacobian and what is the relation between them? What is the relation between the Jacobian of a given transformation and the Jacobian of its inverse? Write a mathematical formula representing this relation. Is the labeling of two coordinate systems (e.g. barred and unbarred) involved in a transformation relation essential or arbitrary? Hence, discuss if the labeling of the coordinates in the Jacobian matrix can be interchanged. Using the transformation equations between the Cartesian and cylindrical coordinate systems, find the Jacobian matrix of the transformation between these systems, i.e. Cartesian to cylindrical and cylindrical to Cartesian. Repeat question 2.32 for the spherical, instead of cylindrical, system to find the Jacobian this time. Give a simple definition of admissible coordinate transformation. What is the meaning of the C n continuity condition? What “invariant” object or property means? Give some illustrating examples. What is the meaning of “composition of transformations”? State a mathematical relation representing such a composition. What is the Jacobian of a composite transformation in terms of the Jacobians of the simple transformations that make the composite transformation? Write a mathematical relation that links all these Jacobians. 2.8 Exercises and Revision 48 2.39 “The collection of all admissible coordinate transformations with non-vanishing Jacobian form a group”. What this means? State your answer in mathematical and descriptive forms. 2.40 Is the transformation of coordinates a commutative operation? Justify your answer by an example. 2.41 A transformation T3 with a Jacobian J3 is a composite transformation, i.e. T3 = T2 T1 where the transformations T1 and T2 have Jacobians J1 and J2 . What is the relation between J1 , J2 and J3 ? 2.42 Two transformations, R1 and R2 , are related by: R1 R2 = I where I is the identity transformation. What is the relation between the Jacobians of R1 and R2 ? What we should call these transformations? 2.43 Discuss the transformation of one set of basis vectors of a given coordinate system to another set of opposite variance type of that system and the relation of this to the metric tensor. 2.44 Discuss the transformation of one set of basis vectors of a given coordinate system to another set of the same variance type of another coordinate system. 2.45 Discuss and compare the results of question 2.43 and question 2.44. Also, compare the mathematical formulation that should apply in each case. 2.46 Define proper and improper coordinate transformations. 2.47 What is the difference between positive and negative orthogonal transformations? 2.48 Give detailed definitions of coordinate curves and coordinate surfaces of 3D spaces discussing the relation between them. 2.49 For each one of the following coordinate systems, what is the shape of the coordinate curves and coordinate surfaces: Cartesian, cylindrical and spherical? 2.50 Make a simple plot representing the φ coordinate curve with the ρ and z coordinate surfaces of a cylindrical coordinate system. 2.51 Make a simple plot representing the r coordinate curve with the θ and φ coordinate surfaces of a spherical coordinate system. 2.52 Define “scale factors” of a coordinate system and outline their significance. 2.53 Give the scale factors of the following coordinate systems: orthonormal Cartesian, cylindrical and spherical. 2.54 Define, mathematically and in words, the covariant and contravariant basis vector sets explaining any symbols involved in these definitions. 2.55 What is the relation of the covariant and contravariant basis vector sets with the coordinate curves and coordinate surfaces of a given coordinate system? Make a simple sketch representing this relation for a general curvilinear coordinate system in a 3D space. 2.56 The covariant and contravariant components of vectors can be transformed one to the other. How? State your answer in a mathematical form. 2.57 What is the significance of the following relations? Ei · Ej = δij Ei · Ej = δ ij 2.58 Write down the mathematical relations that correlate the basis vectors to the compo- 2.8 Exercises and Revision 49 nents of the metric tensor in their covariant and contravariant forms. 2.59 Using Eq. 50, show that if E1 , E2 , E3 is a right handed orthonormal system then Ei = Ei . Repeat the question using this time Eq. 51 where E1 , E2 , E3 form a right handed orthonormal system. Hence, conclude that when the covariant or contravariant basis vector set is orthonormal then the covariant and contravariant components of a given tensor are identical. 2.60 State the mathematical relations between the original and transformed (i.e. unbarred and barred) basis vector sets in their covariant and contravariant forms under admissible coordinate transformations. 2.61 Correct, if necessary, the following equations explaining all the symbols involved: 1 E1 · (E2 × E3 ) = √ g  √ E1 · E2 × E3 = g 2.62 Obtain the relation: g = J 2 from the relation: JT J = [gij ] giving full explanation of each step. 2.63 State three consequences of having mutually orthogonal contravariant basis vectors at each point in the space justifying these consequences. 2.64 Discuss the relationship between the concepts of space, coordinate system and metric tensor. Chapter 3 Tensors The subject of this chapter is tensors in general. The chapter is divided into three sections where we discuss in these sections: tensor types, tensor operations, and tensor representations. 3.1 Tensor Types In the following subsections, we introduce a number of tensor types and categories and highlight their main characteristics and differences. These types and categories are not mutually exclusive and hence they overlap in general. Moreover, they may not be exhaustive in their classes as some tensors may not instantiate any one of a complementary set of types such as being symmetric or anti-symmetric. 3.1.1 Covariant and Contravariant Tensors These are the main types of tensor with regard to the rules of their transformation between different coordinate bases. Covariant tensors are notated with subscript indices (e.g. Ai ) while contravariant tensors are notated with superscript indices (e.g. Aij ). A covariant tensor is transformed according to the following rule: ∂uj Aj ∂ ūi while a contravariant tensor is transformed according to the following rule: Āi = (71) ∂ ūi j A (72) ∂uj The barred and unbarred symbols in these equations represent the same mathematical object (tensor or coordinate) in the transformed and original coordinate systems respectively. An example of covariant tensors is the gradient of a scalar field while an example of contravariant tensors is the displacement vector. The above transformation rules, as demonstrated by Eqs. 71 and 72, which correspond to rank-1 tensors can be easily extended to tensors of any rank following the above pattern. Hence, a covariant tensor A of rank m is transformed according to the following rule: Āi = ∂up ∂uq ∂ur · · · Apq···r (73) ∂ ūi ∂ ūj ∂ ūm Similarly, a contravariant tensor B of rank n is transformed according to the following rule: ∂ ūn pq···r ∂ ūi ∂ ūj · · · B (74) B̄ ij···n = ∂up ∂uq ∂ur Āij···m = 50 3.1.1 Covariant and Contravariant Tensors 51 Some tensors of rank > 1 have mixed variance type, i.e. they are covariant in some indices and contravariant in others. In this case the covariant variables are indexed with subscripts while the contravariant variables are indexed with superscripts, e.g. Aji which is covariant in i and contravariant in j. A mixed type tensor transforms covariantly in its covariant indices and contravariantly in its contravariant indices, e.g. ∂ ūl ∂uj ∂ ūn i k A (75) ∂ui ∂ ūm ∂uk j In brief, tensors of rank-n (n > 0) whose all free indices are subscripts are covariant, and tensors of rank-n (n > 0) whose all free indices are superscripts are contravariant, while tensors of rank-n (n > 1) whose some of free indices are subscripts while other free indices are superscripts are mixed. Hence, Ai and Bij are covariant and transform according to Eq. 73, Ai and B ijk are contravariant and transform according to Eq. 74, and Aji and Bijk are mixed and transform according to Eq. 75. As indicated before, for orthonormal Cartesian coordinate systems there is no difference between the covariant and contravariant representations, and hence the indices can be upper or lower although it is common to use lower indices in this case. The practical rules for writing the transformation equations, as seen in the above equations, can be summarized in the following points where we take Eq. 75 (which is sufficiently general) as an example to illustrate the steps taken in writing the transformation equations: 1. We start by writing the symbol of the transformed tensor on the left hand side of the transformation equation and the symbol of the original tensor on the right hand side, that is: Ā = A (76) Ālmn = 2. We index the transformed tensor according to its intended indicial structure observing the order of the indices. We similarly index the original tensor with its original indices, that is: Ālmn = Ai jk (77) 3. We insert a number of partial differential operators on the right hand side equal to the number of free indices, that is: ∂u ∂u ∂u i k A (78) ∂u ∂u ∂u j 4. We index the coordinates of the transformed tensor in the numerator or denominator in these operators according to the order of the indices in the tensor where these indices are in the same position (upper or lower) as their position in the tensor, that is: ∂ul ∂u ∂un i k A (79) Ālmn = ∂u ∂um ∂u j 5. Because the transformed tensor is barred then its coordinates should also be barred, that is: ∂ ūl ∂u ∂ ūn i k Ālmn = A (80) ∂u ∂ ūm ∂u j Ālmn = 3.1.1 Covariant and Contravariant Tensors 52 6. We then index the coordinates of the original tensor in the numerator or denominator in these operators according to the order of the indices in the tensor where these indices are in the opposite position (upper or lower) to their position in the tensor, that is: ∂ ūl ∂uj ∂ ūn i k l n Ā m = A (81) ∂ui ∂ ūm ∂uk j We note that an upper coordinate index in the denominator of the partial differential operators is equivalent to a lower tensor index. We also note that each pair of the transformed indices are on the two sides and hence they should be in the same position (upper or lower) according to the rules of free indices (see § 1.2), while the pairs of indices of the original tensor are in the same side and hence they should be in opposite positions according to the rules of dummy indices. The covariant and contravariant types of a tensor are linked through the metric tensor. As will be detailed later (refer to § 4.5), the contravariant metric tensor is used for raising covariant indices of covariant and mixed tensors, e.g. Ai = g ik Ak Aij = g ik Akj (82) where the contravariant metric tensor g ik is used to raise the covariant index k to become the contravariant index i. Similarly, the covariant metric tensor is used for lowering contravariant indices of contravariant and mixed tensors, e.g. Ai = gik Ak Aij = gik Akj (83) where the covariant metric tensor gik is used to lower the contravariant index k to become the covariant index i. Hence, the metric tensor is used to change the variance type of tensor indices and for that reason it is labeled as index raising or index lowering or index shifting operator. We note that in the raising and lowering operations, the metric tensor acts as an index replacement operator, as well as a shifting operator, by changing the label of the shifted index, as seen in the above examples where the shifted index k is replaced with i which is the other index of the metric tensor. Because it is possible to shift the index position of a tensor by using the covariant and contravariant types of the metric tensor as an index shifting operator, a given tensor can be cast into a covariant or a contravariant form, as well as a mixed form in the case of tensors of rank > 1. However, it should be emphasized that the order of the indices must be respected in this process, because two tensors with the same indicial structure but with different indicial order are not equal in general, as stated before. For example: Aij = gjk Aik 6= gjk Aki = Aj i (84) Dots may be inserted to remove any ambiguity about the order of the indices and hence Ai· j or Aij. means i first and j second while A·ji or Aji. means j first and i second. A tensor of m free contravariant indices and n free covariant indices may be called type (m, n) tensor. When one or both variance types are absent, zero is used to refer to the absent variance type in this notation. Accordingly, Akij is a type (1, 2) tensor, B ik is a 3.1.1 Covariant and Contravariant Tensors 53 ts type (2, 0) tensor, Cm is a type (0, 1) tensor, and Dpqr is a type (2, 3) tensor. As we will see in § 3.1.3, the type may also include the weight as a third entry and hence the type in this sense is identified by the symbol (m, n, w) where m and n refer to the number of contravariant and covariant indices respectively while w refers to the weight of the tensor. It is obvious that the rank of the tensor can be obtained from this notation by adding the first two entries and hence a tensor of type (m, n) or (m, n, w) is of rank m + n. As seen earlier, the vectors providing the basis set for a coordinate system are of covariant type when they are tangent to the coordinate curves, which represent the curves along which exactly one coordinate varies while the other coordinates are held constant, and they are of contravariant type when they are perpendicular to the coordinate surfaces which are the surfaces over which all coordinates vary except one which is held constant. Formally, the covariant and contravariant basis vectors are given respectively by: Ei = ∂r ∂ui Ei = ∇ui (85) where r = xj ej is the position vector in Cartesian coordinates and ui is a general coordinate. As indicated earlier, a superscript in the denominator of partial derivatives is equivalent to a subscript in the numerator and hence the above equation is consistent with the rules of tensor indices which were outlined in § 1.2. We also remark that in general the basis vectors, whether covariant or contravariant, are not necessarily of unit length and/or mutually orthogonal although they may be so. In fact, there are standard mathematical procedures to orthonormalize the basis set if it is not and if orthonormalization is needed. For example, the two covariant basis vectors, E1 and E2 , of a 2D space (i.e. surface) can be orthonormalized as follows: E1 = E1 E1 =√ |E1 | g11 E2 = g11 E2 − g12 E1 √ g11 g (86) where g is the determinant of the surface covariant metric tensor, the indexed g are the coefficients of this tensor, and the underlined vectors are orthonormal basis vectors, that is: E2 · E2 = 1 E1 · E2 = 0 (87) E1 · E1 = 1 This can be verified by conducting the dot products of the last equation using the vectors defined in Eq. 86. As seen earlier, the two sets of covariant and contravariant basis vectors are reciprocal systems and hence they satisfy the following reciprocity relations: Ei · Ej = δij Ei · Ej = δ ij (88) where the indexed δ is the Kronecker delta tensor (refer to § 4.1) which can be represented by the unity matrix. The reciprocity of these two sets of basis vectors is illustrated schematically in Figure 11 for the case of a 2D space. 3.1.1 Covariant and Contravariant Tensors 54 E2 E2 φ E1 φ E1 Figure 11: The reciprocity relation between the covariant and contravariant basis vectors in a 2D space where E1 ⊥ E2 , E1 ⊥ E2 , and |E1 | |E1 | cos φ = |E2 | |E2 | cos φ = 1. A vector can be represented either by covariant components with contravariant basis vectors or by contravariant components with covariant basis vectors. For example, a vector A can be expressed as: A = Ai Ei or A = Ai Ei (89) where Ei and Ei are the contravariant and covariant basis vectors respectively. This is illustrated graphically in Figure 12 for a vector A in a 2D space. The use of the covariant or contravariant form of the vector representation is a matter of choice and convenience since these two representations are equivalent as they represent and correctly describe the same object. More generally, a tensor of any rank (≥ 1) can be represented covariantly using contravariant basis tensors of that rank, or contravariantly using covariant basis tensors, or in a mixed form using a mixed basis of opposite type. For example, a rank-2 tensor A can be written as: A = Aij Ei Ej = Aij Ei Ej = Aij Ei Ej = Aij Ei Ej (90) where Ei Ej , Ei Ej , Ei Ej and Ei Ej are dyadic products of the basis vectors of the presumed system (refer to § 3.2.3). We remark that dyadic products represent a combination of two vectors and hence they represent two directions in a certain order. Figure 13 is a graphic illustration of the nine dyadic products of the three unit basis vectors in a 3D space with a rectangular Cartesian coordinate system. Similarly, a rank-3 tensor B can be written as: B = Bijk Ei Ej Ek = B ijk Ei Ej Ek = Bi jk Ei Ej Ek = B i j k Ei Ej Ek = · · · etc. (91) where Ei Ej Ek . . . etc. are triads. More generally, a rank-n tensor C can be written as: C = Cij···n Ei Ej · · · En = C ij···n Ei Ej · · · En = Ci j···n Ei Ej · · · En = · · · etc. (92) 3.1.2 True and Pseudo Tensors 55 A2 |E2| A E2 A2 E2 E2 E1 A1 E1 E1 A1 |E1| Figure 12: The representation of a vector A in covariant and contravariant basis vector sets in a 2D space where the components shown at the four points are with reference to unit vectors in the given directions, e.g. A1 |E1 | is a component with reference to a unit vector in the direction of E1 . where Ei Ej · · · En . . . etc. are n-polyads. We note that the order of the indices in the tensor bases (i.e. dyads, triads and n-polyads) should be the same as the order of the indices in the tensor components, as seen in the above examples. Finally, it should be remarked that the two sets of basis vectors (i.e. covariant and contravariant), like the components of the tensors themselves, are identical for orthonormal Cartesian systems. This can be explained by the fact that since the coordinate surfaces are mutually perpendicular planes then the coordinate curves are mutually perpendicular straight lines and hence the tangent vectors to the xi coordinate curve is the same as the perpendicular vector to the xi coordinate surface noting that the coordinates are scaled uniformly in all directions. 3.1.2 True and Pseudo Tensors These are also called polar and axial tensors respectively although it is more common to use these terms for vectors. Pseudo tensors may also be called tensor densities, however the terminology in this part, like many other parts, is not universal. True tensors are proper or ordinary tensors and hence they are invariant under coordinate transformations, while pseudo tensors are not proper tensors since they do not transform invariantly as they acquire a minus sign under improper orthogonal transformations which involve inversion of coordinate axes through the origin of coordinates with a change of system handedness. Figure 14 demonstrates the behavior of a true vector v and a pseudo vector p where the former keeps its direction in the space following a reflection of the coordinate system through the origin of coordinates while the latter reverses its direction following 3.1.2 True and Pseudo Tensors 56 x3 x1 e1 e1 x3 x2 x1 x3 x1 e2 e1 e3 e1 x2 x1 x3 x2 x1 x3 x1 e1 e2 x3 e2 e2 x1 e3 e2 x2 x3 x2 x1 x3 x2 e1 e3 e2 e3 x2 x3 x2 x1 e3 e3 x2 Figure 13: The nine unit dyads associated with the double directions of rank-2 tensors in a 3D space with a rectangular Cartesian coordinate system. The vectors ei and ej (i, j = 1, 2, 3) are unit vectors in the directions of coordinate axes where the first indexed e (blue) represents the first vector of the dyad while the second indexed e (red) represents the second vector of the dyad. In these nine frames, the first vector is fixed along each row while the second vector is fixed along each column. this operation. Because true and pseudo tensors have different mathematical properties and represent different types of physical entities, the terms of consistent tensor expressions and equations should be uniform in their true and pseudo type, i.e. all terms should be true or all terms should be pseudo. The direct product (refer to § 3.2.3) of true tensors is a true tensor. The direct product of even number of pseudo tensors is a true tensor, while the direct product of odd number of pseudo tensors is a pseudo tensor. The direct product of a mix of true and pseudo tensors is a true or pseudo tensor depending on the number of pseudo tensors involved in the product as being even or odd respectively. Similar rules to those of the direct product apply to the cross product, including the curl operation, involving tensors (which are usually of rank-1) with the addition of a pseudo 3.1.3 Absolute and Relative Tensors 57 x3 v X2 V X1 p O O P x2 x1 X3 Figure 14: The behavior of a true vector (v and V) and a pseudo vector (p and P) on reflecting the coordinate system in the origin of coordinates. The lower case symbols stand for the objects in the original system while the upper case symbols stand for the same objects in the reflected system. factor for each cross product operation. This factor is contributed by the permutation tensor which is implicit in the definition of the cross product. As we will see in § 4.2, the permutation tensor is a pseudo tensor. In summary, what determines the tensor type (true or pseudo) of the tensor terms involving direct and cross products is the parity of the multiplicative factors of pseudo type plus the number of cross product operations involved since each cross product operation contributes a permutation tensor. We note that inner product (see § 3.2.5) is the result of a direct product (see § 3.2.3) operation followed by a contraction (see § 3.2.4) and hence it is like a direct product in this context. 3.1.3 Absolute and Relative Tensors Considering an arbitrary transformation from a general coordinate system to another, a tensor of weight w is defined by the following general tensor transformation relation: Āij...klm...n ∂u = ∂ ū w ∂ ūk ∂ud ∂ue ∂uf ab...c ∂ ūi ∂ ūj · · · · · · A de...f ∂ua ∂ub ∂uc ∂ ūl ∂ ūm ∂ ūn (93) where ∂u symbolizes the Jacobian of the transformation between the two systems (see § ∂ ū 2.3). When w = 0 the tensor is described as an absolute or true tensor, and when w 6= 0 the tensor is described as a relative tensor. When w = −1 the tensor may be described as a pseudo tensor, while when w = 1 the tensor may be described as a tensor density. We note that some of these labels are used differently by different authors as the terminology of tensor calculus is not universally approved and hence the conventions of each author should 3.1.4 Isotropic and Anisotropic Tensors 58 be checked. Also, there is an obvious overlap between this classification (i.e. absolute and relative) and the previous classification (i.e. true and pseudo) at least according to some conventions. As indicated earlier (see § 3.1.1), a tensor of m contravariant indices and n covariant indices may be described as a tensor of type (m, n). This may be extended to include the weight w as a third entry and hence the type of the tensor is identified by (m, n, w). Relative tensors can be added and subtracted (see § 3.2.1) if they have the same indicial structure and have the same weight; the result is a tensor of the same indicial structure and weight. Also, relative tensors can be equated if they have the same indicial structure and weight. In brief, the terms of tensor expressions and equalities should have the same weight w, whether w is equal to zero or not. Multiplication of relative tensors produces a relative tensor whose weight is the sum of the weights of the original tensors. Hence, if the weights are added up to a non-zero value the result is a relative tensor of that weight; otherwise it is an absolute tensor. We remark that the statements of the previous paragraph can be generalized by including w = 0 which corresponds to absolute tensors and hence “relative” in those statements is more general than being opposite to “absolute”. Accordingly, and from the perspective of relative tensors (i.e. assuming that other qualifications such as matching in indicial structure, are met), two absolute tensors can be added or subtracted or equated but an absolute and a relative tensor (i.e. with w 6= 0) cannot since they are “relative” tensors with different weights. 3.1.4 Isotropic and Anisotropic Tensors Isotropic tensors are characterized by the property that the values of their components are invariant under coordinate transformation by proper rotation of axes. In contrast, the values of the components of anisotropic tensors are dependent on the orientation of the coordinate axes. Notable examples of isotropic tensors are scalars (rank-0), the vector 0 (rank-1), Kronecker delta (rank-2) and Levi-Civita tensor (rank ≥ 2). Many tensors describing physical properties of materials, such as stress and magnetic susceptibility, are anisotropic. Direct and inner products (see § 3.2.3 and 3.2.5) of isotropic tensors are isotropic tensors. The zero tensor of any rank and any dimension is isotropic; therefore if the components of a tensor vanish in a particular coordinate system they will vanish in all properly and improperly rotated coordinate systems.[14] Consequently, if the components of two tensors are identical in a particular coordinate system they are identical in all transformed coordinate systems since the tensor of their difference is a zero tensor and hence it is invariant. This means that tensor equalities and identities are invariant under coordinate transformations, which is one of the main motivations for the use of tensors in mathematics and science. As indicated, all rank-0 tensors (scalars) are isotropic. Also, the zero vector, 0, of any dimension is isotropic; in fact it is the only rank-1 isotropic tensor. [14] For improper rotation, this is more general than being isotropic. 3.1.5 Symmetric and Anti-symmetric Tensors 3.1.5 59 Symmetric and Anti-symmetric Tensors These types of tensor apply to high ranks only (rank ≥ 2) since symmetry and antisymmetry of tensors require in their definition two free indices at least, and hence a scalar with no index and a vector with a single index do not qualify to be symmetric or anti-symmetric. Moreover, these types are not exhaustive, even for tensors of rank ≥ 2, as there are high-rank tensors which are neither symmetric nor anti-symmetric, and hence we may call them asymmetric tensors although this terminology may include anti-symmetric tensors as well. A rank-2 tensor Aij is symmetric in its components iff for all i and j the following condition is satisfied: Aji = Aij (94) and anti-symmetric or skew-symmetric iff for all i and j the following condition is satisfied: Aji = −Aij (95) Similar conditions apply to contravariant type tensors, i.e. Aji = Aij for symmetric and Aji = −Aij for anti-symmetric. This also applies to the following definitions and identities which are largely presented in covariant forms. More generally, a rank-n tensor Ai1 ...in is symmetric in its two indices ij and il iff the following condition applies identically: Ai1 ...il ...ij ...in = Ai1 ...ij ...il ...in (96) and anti-symmetric in its two indices ij and il iff the following condition applies identically: Ai1 ...il ...ij ...in = −Ai1 ...ij ...il ...in (97) Any rank-2 tensor Aij can be synthesized from (or decomposed into) a symmetric part A(ij) , which is marked with round brackets enclosing the indices, and an anti-symmetric part A[ij] , which is marked with square brackets, where the following relations apply: Aij = A(ij) + A[ij] 1 (Aij + Aji ) A(ij) = 2 1 (Aij − Aji ) A[ij] = 2 (98) (99) (100) The first relation can be verified by substituting the second and third relations into the first, that is: Aij = A(ij) + A[ij] = 1 1 1 (Aij + Aji ) + (Aij − Aji ) = (2Aij ) = Aij 2 2 2 (101) which is an identity, while the second and third relations can be verified by shifting the indices, that is: A(ji) = 1 1 (Aji + Aij ) = (Aij + Aji ) = A(ij) 2 2 (102) 3.1.5 Symmetric and Anti-symmetric Tensors A[ji] = 1 1 (Aji − Aij ) = − (Aij − Aji ) = −A[ij] 2 2 60 (103) and hence A(ij) is symmetric and A[ij] is anti-symmetric. Similarly, a rank-3 tensor Aijk can be symmetrized by the following relation: A(ijk) = 1 (Aijk + Akij + Ajki + Aikj + Ajik + Akji ) 3! (104) and anti-symmetrized by the following relation: A[ijk] = 1 (Aijk + Akij + Ajki − Aikj − Ajik − Akji ) 3! (105) where the first three terms in these equations are the even permutations of the indices ijk and the last three terms are the odd permutations of these indices. These relations can also be verified by exchanging the indices. For example: A(jik) = 1 (Ajik + Akji + Aikj + Ajki + Aijk + Akij ) = A(ijk) 3! (106) and hence it is symmetric in i and j. Similarly: A[jik] = 1 (Ajik + Akji + Aikj − Ajki − Aijk − Akij ) = −A[ijk] 3! (107) and hence it is anti-symmetric in i and j. The exchange of other indices can be done similarly and hence the symmetric nature of A(ijk) and the anti-symmetric nature of A[ijk] with respect to their other indices can also be verified. The symmetric and anti-symmetric parts, A(ijk) and A[jik] , are related by the following equations:  (108) 3 A(ijk) + A[jik] = Aijk + Akij + Ajki  3 A(ijk) − A[jik] = Aikj + Ajik + Akji (109) which can also be verified by substitution, as done for rank-2. The right hand side of the first equation is the sum of the even permutations of the indices ijk while the right hand side of the second equation is the sum of the odd permutations of these indices. More generally, a rank-n tensor Ai1 ...in can be symmetrized by: A(i1 ...in ) =  X 1 X even permutations of i’s + odd permutations of i’s n! and anti-symmetrized by:  X 1 X A[i1 ...in ] = even permutations of i’s − odd permutations of i’s n! (110) (111) Similar verifications and relations can be established for the rank-n case following the patterns seen in the above rank-2 and rank-3 cases. 3.1.5 Symmetric and Anti-symmetric Tensors 61 A tensor of high rank (> 2) may be symmetrized or anti-symmetrized with respect to only some of its indices instead of all of its indices. For example, in the following the tensor Aijk is symmetrized and anti-symmetrized respectively with respect to its first two indices only, that is: 1 (Aijk + Ajik ) 2 1 (Aijk − Ajik ) = 2 A(ij)k = (112) A[ij]k (113) The symmetry of A(ij)k and the anti-symmetry of A[ij]k with respect to the ij indices can be verified by exchanging these indices in the above relations as done previously for A(ij) and A[ij] . Moreover, the tensor Aijk can be expressed as the sum of these symmetric and anti-symmetric tensors, that is: A(ij)k + A[ij]k = 1 1 (Aijk + Ajik ) + (Aijk − Ajik ) = Aijk 2 2 (114) A tensor is described as totally symmetric iff it is symmetric with respect to all of its indices, that is: Ai1 ...in = A(i1 ...in ) (115) and hence its anti-symmetric part is zero, i.e. A[i1 ...in ] = 0. Similarly, a tensor is described as totally anti-symmetric iff it is anti-symmetric in all of its indices, that is: Ai1 ...in = A[i1 ...in ] (116) and hence its symmetric part is zero, i.e. A(i1 ...in ) = 0. For a totally anti-symmetric tensor, non-zero entries can occur only when all the indices are different. This is because an exchange of two identical indices, which identifies identical entries, should change the sign due to the anti-symmetry and no number can be equal to its negation except zero. Therefore, if the tensor Aijk is totally anti-symmetric then only its entries of the form Aijk , where i 6= j 6= k, are not identically zero while all the other entries (i.e. all those of the from Aiij , Aiji , Ajii and Aiii where i 6= j with no sum on i) vanish identically. In fact, this also applies to partially anti-symmetric tensors where the entries corresponding to identical anti-symmetric indices should vanish identically. Hence, if Aijkl is anti-symmetric in its first two indices and in its last two indices then only its entries with i 6= j and k 6= l do not vanish identically while all its entries with i = j or k = l (or both) are identically zero. It should be remarked that the indices whose exchange defines the symmetry and antisymmetry relations should be of the same variance type, i.e. both upper or both lower. Hence, a tensor Aji is not symmetric if the components satisfy the relation Aji = Aij or anti-symmetric if the components satisfy the relation Aji = −Aij . This should be obvious by considering that the covariant and contravariant indices correspond to different basis sets (i.e. contravariant and covariant), as explained in § 3.1.1. Another important remark is that the symmetry and anti-symmetry characteristic of a tensor is invariant under coordinate transformations. Hence, a symmetric/anti-symmetric 3.1.6 General and Affine Tensors 62 tensor in one coordinate system is symmetric/anti-symmetric in all other coordinate systems. This is because if we label the tensor as A and its transformation by symmetry or anti-symmetry as B, then an equality of the two tensors based on their symmetric or anti-symmetric nature (i.e. Aij = Bji or Aij = −Bji ) can be expressed, using an algebraic transformation, as a zero tensor (i.e. Aij − Bji = 0 or Aij + Bji = 0) which is invariant under coordinate transformations as stated earlier (see § 3.1.4). Similarly, a tensor which is neither symmetric nor anti-symmetric in one coordinate system should remain so in all other coordinate systems obtained by permissible transformations. This is based on the previous statement because if it is symmetric or anti-symmetric in one coordinate system it should be symmetric or anti-symmetric in all other coordinate systems according to the previous statement. In brief, the characteristic of a tensor as being symmetric or anti-symmetric or neither is invariant under permissible coordinate transformations. Finally, for a symmetric tensor Aij and an anti-symmetric tensor B ij (or the other way around) we have the following useful and widely used identity: Aij B ij = 0 (117) This is because an exchange of the i and j indices will change the sign of one tensor only, which is the anti-symmetric tensor, and this will change the sign of the term in the summation resulting in having a sum of terms which is identically zero due to the fact that each term in the sum has its own negation. This also includes the zero components of the anti-symmetric tensor where the terms containing these components are zero. 3.1.6 General and Affine Tensors Affine tensors are tensors that correspond to admissible linear coordinate transformations (see § 2.3.4) from an original rectangular system of coordinates, while general tensors correspond to any type of admissible coordinate transformations. These categories are part of the terminology of tensor calculus which we use in this book and hence they do not have a particular significance. 3.2 Tensor Operations In this section, we briefly examine the main elementary tensor operations, which are mostly of algebraic nature, that permeate tensor algebra and calculus. There are various operations that can be performed on tensors to produce other tensors in general. Examples of these operations are addition/subtraction, multiplication by a scalar (rank-0 tensor), multiplication of tensors (each of rank > 0), contraction and permutation. Some of these operations, such as addition and multiplication, involve more than one tensor while others, such as permutation, are performed on a single tensor. Contraction can involve one tensor or two tensors. Before we start our investigation, we should remark that the last subsection (i.e. § 3.2.7), which is about the quotient rule of tensors, is added to the present section of tensor operations because this section is the most appropriate place for it in the present 3.2.1 Addition and Subtraction 63 book considering the dependency of the definition of this rule on other tensor operations; otherwise the subsection is not about a tensor operation in the same sense as the operations presented in the other subsections. Another remark is that in tensor algebra division is allowed only for scalars, and hence if the components of an indexed tensor should appear in a denominator, the tensor should be redefined to avoid this. For example, if Ai represents the components of a tensor and we should have A1i as the components of another tensor, then we introduce another tensor, say Bi , whose components are the reciprocals of Ai (i.e. Bi = A1i ) and use Bi instead of A1i . The purpose of this change of notation is to avoid confusion and facilitate the manipulation of tensors according to the familiar rules of indices. 3.2.1 Addition and Subtraction Tensors of the same rank and type can be added algebraically to produce a tensor of the same rank and type, e.g. a=b+c Ai = B i − C i Aij = Bji + Cji (118) In this operation the entries of the two tensors are added algebraically componentwise. For example, in the second equality of the last equation the C1 component of the tensor C is subtracted from the B1 component of the tensor B to produce the A1 component of the tensor A, while in the third equality the B21 component of the tensor B is added to the C21 component of the tensor C to produce the A12 component of the tensor A. We note that “type” in the above statement refers to variance type (covariant, contravariant, mixed) and true/pseudo type as well as other qualifications to which the tensors participating in an addition or subtraction operation should match such as having the same weight if they are relative tensors, as outlined previously (refer for example to § 3.1.3). The added/subtracted terms should also have the same indicial structure with regard to their free indices, as explained in § 1.2 in the context of the rules governing the indices j of tensor expressions and equalities. Hence, the tensors Aijk and Bik cannot be added or i subtracted although they are of the same rank and type, but the tensors Aijk and Bjk can be added and subtracted. In brief, the tensors involved in addition and subtraction operations should satisfy all the rules that have been stated previously on the terms of tensor expressions and equalities. Addition of tensors is associative and commutative, that is: (A + B) + C = A + (B + C) A+B = B+A (119) (120) In fact, associativity and commutativity can include subtraction if the minus sign is absorbed in the subtracted tensor; in which case the operation is converted to addition. 3.2.2 Multiplication of Tensor by Scalar 3.2.2 64 Multiplication of Tensor by Scalar A tensor can be multiplied by a scalar, which should not be zero in general, to produce a tensor of the same variance type, rank and indicial structure, e.g. j Ajik = aBik (121) where a is a non-zero scalar. As indicated by this equation, multiplying a tensor by a scalar means multiplying each component of the tensor by that scalar. Multiplication by a scalar is commutative and distributive over algebraic sum of tensors, that is: aA = Aa a (A ± B) = aA ± aB (122) (123) where a is a scalar and A and B are non-scalar tensors. It is also associative when more than two factors are involved, that is: a (bA) = (ab) A a (A ◦ B) = (aA) ◦ B = A ◦ (aB) (124) (125) where b is another scalar and ◦ represents an outer or inner product operation (see § 3.2.3 and 3.2.5). 3.2.3 Tensor Multiplication This operation, which can be defined generically as multiplication between two non-scalar tensors, may also be called outer or exterior or direct or dyadic multiplication, although some of these names may be reserved to operations on vectors. On multiplying each component of a tensor of rank r by each component of a tensor of rank k, both of dimension d, a tensor of rank (r+k) with dr+k components is obtained where the variance type of each index (covariant or contravariant) is preserved. More clearly, by multiplying a tensor of type (m1 , n1 , w1 ) by a tensor of type (m2 , n2 , w2 ) a tensor of type (m1 +m2 , n1 +n2 , w1 +w2 ) is obtained. This means that the tensor rank and weight in the outer product operation are additive and the operation conserves the variance type of each index of the tensors involved. Also, the order of the indices in the product should match the order of the indices in the multiplicands, as will be seen in the following examples, since tensor indices refer to specific tensor bases and the multiplication of tensors is not commutative (refer to § 3.1.1). For example, if A and B are covariant tensors of rank-1, then on multiplying A by B we obtain a covariant tensor C of rank-2 and type (0, 2) where the components of C are given by: Cij = Ai Bj (126) while on multiplying B by A we obtain a covariant tensor D of rank-2 and type (0, 2) where the components of D are given by: Dij = Bi Aj (127) 3.2.3 Tensor Multiplication 65 Similarly, if A is a contravariant tensor of rank-2 and B is a covariant tensor of rank-2, then on multiplying A by B we obtain a mixed tensor C of rank-4 and type (2, 2) where the components of C are given by: C ijkl = Aij Bkl (128) while on multiplying B by A we obtain a mixed tensor D of rank-4 and type (2, 2) where the components of D are given by: Dijkl = Bij Akl (129) In general, the outer product of tensors yields a tensor. In the outer product operation, it is generally understood that all the indices of the involved tensors have the same range (i.e. all the tensors have the same dimension) although this may not always be the case. As indicated before (see § 1.2), there are cases of tensors which are not uniformly dimensioned, and in some cases these tensors may be regarded as the result of an outer product of lower rank tensors. The direct multiplication of tensors may be marked by the symbol , mostly when using symbolic notation for tensors, e.g. A  B. However, in the present book no symbol will be used to represent the operation of direct multiplication[15] and hence the operation is symbolized by putting the symbols of the tensors side by side, e.g. AB where A and B are non-scalar tensors. In this regard, the reader should be vigilant to avoid confusion with the operation of matrix multiplication which, according to the notation of matrix algebra, is also symbolized as AB where A and B are matrices of compatible dimensions, since matrix multiplication from tensor perspective is an inner product, rather than an outer product, operation. The direct multiplication of tensors is not commutative in general as indicated above; however it is distributive with respect to algebraic sum of tensors, that is: (130) AB 6= BA A (B ± C) = AB ± AC (B ± C) A = BA ± CA (131) Regarding the associativity of direct multiplication, there are cases in which this operation is not associative according to the research literature of tensor calculus. As indicated before, the rank-2 tensor constructed by the direct multiplication of two vectors is commonly called dyad. More generally, tensors may be expressed as an outer product of vectors where the rank of the resultant product is equal to the number of the vectors involved, i.e. 2 for dyads, 3 for triads and so on. However, not every tensor can be synthesized as a product of lower rank tensors. Multiplication of a tensor by a scalar (refer to § 3.2.2) may be regarded as a special case of direct multiplication since it is a tensor multiplication operation with one of the tensors involved being of rank-0 although the definition of direct multiplication seems to restrict this operation to non-scalar tensors, as stated above. [15] We mean “specifically” because we use the symbol ◦ to represent general tensor multiplication which includes inner and outer tensor products. 3.2.4 Contraction 3.2.4 66 Contraction The contraction operation of a tensor of rank > 1 is to make two free indices identical, by unifying their symbols, followed by performing summation over these repeated indices, e.g. Aji contraction Aii (132) −−−−−−−−→ contraction on jl Amk Ajk (133) im il −−−−−−−−−−−−→ Accordingly, contraction results in a reduction of the rank by 2 since it requires the annihilation of two free indices by the summation operation. Therefore, the contraction of a rank-2 tensor results in a scalar, the contraction of a rank-3 tensor results in a vector, the contraction of a rank-4 tensor results in a rank-2 tensor, and so on. Contraction can also take place between two tensors as part of an inner product operation, as will be explained in § 3.2.5, although the contraction in this case is actually done on the tensor resulting from the outer product operation that underlies the inner product. For general coordinate systems, the pair of contracted indices should be different in their variance type, i.e. one upper and one lower. Hence, contraction of a mixed tensor of type (m, n) will in general produce a tensor of type (m − 1, n − 1). A tensor of type (p, q) can therefore have p × q possible contractions, i.e. one contraction for each combination of lower and upper indices. As indicated before, there is no difference between the covariant and contravariant types in orthonormal Cartesian systems and hence contraction can take place between any pair of indices. Accordingly, a rank-n tensor in orthonormal Cartesian possible individual contraction operations where each one of the n systems can have n(n−1) 2 indices can be contracted with each one of the remaining (n − 1) indices and the factor 2 in the denominator represents the fact that the contraction operation is independent of the order of the two contracted indices since contracting i with j is the same as contracting j with i.[16] We note that conducting a contraction operation on a tensor results into a tensor. Similarly, the application of a contraction operation on a relative tensor (see § 3.1.3) produces a relative tensor of the same weight as the original tensor. A common example of a contraction operation conducted on a single tensor is the operation of taking the trace of a square matrix, as defined in matrix algebra, by summing its diagonal elements, which can be considered as a contraction operation on the rank-2 tensor represented by the matrix, and hence it yields the trace which is a scalar. Similarly, a well known example of a contraction operation that takes place between two tensors is the dot product operation on vectors which can be considered as a direct multiplication (refer to § 3.2.3) of the two vectors that results in a rank-2 tensor followed by a contraction operation and hence it produces a scalar. 3.2.5 Inner Product On taking the outer product (refer to § 3.2.3) of two tensors of rank ≥ 1 followed by a contraction (refer to § 3.2.4) on two indices of the product, an inner product of the two [16] This may also be formulated by the rule of combination. 3.2.5 Inner Product 67 tensors is formed. Hence, if one of the original tensors is of rank-m and the other is of rank-n, the inner product will be of rank-(m + n − 2). In the symbolic notation of tensor calculus, the inner product operation is usually symbolized by a single dot between the two tensors, e.g. A · B, to indicate the contraction operation which follows the outer multiplication. In general, the inner product is not commutative. When one or both of the tensors involved in the inner product are of rank > 1 then the order of the multiplicands does matter in general, that is: A · B 6= B · A (134) However, the inner product operation is distributive with respect to the algebraic sum of tensors, that is: A · (B ± C) = A · B ± A · C (B ± C) · A = B · A ± C · A (135) As indicated before (see § 3.2.4), the dot product of two vectors is an example of the inner product of tensors, i.e. it is an inner product of two rank-1 tensors to produce a rank-0 tensor. For example, if a is a covariant vector and b is a contravariant vector, then their dot product can be depicted as follows: [ab]ij = ai bj contraction −−−−−−−−→ a · b = ai bi (136) Another common example, from linear algebra, of inner product is the multiplication of a matrix representing a rank-2 tensor, by a vector, which is a rank-1 tensor, to produce a vector. For example, if A is a rank-2 covariant tensor and b is a contravariant vector, then their inner product can be depicted, according to tensor calculus, as follows: [Ab]ijk = Aij bk contraction on jk −−−−−−−−−−−−−→ [A · b]i = Aij bj (137) This operation is equivalent to the aforementioned operation of multiplying a matrix by a vector as defined in linear algebra. We note that we are using here the symbolic notation of tensor calculus, rather than the matrix notation, in writing Ab and A · b to represent, respectively, the outer and inner products. In matrix notation, Ab is used to represent the product of a matrix by a vector which is an inner product according to the terminology of tensor calculus. The multiplication of two n × n matrices, as defined in linear algebra, to produce another n × n matrix is another example of inner product. In this operation, each one of the matrices involved in the multiplication, as well as the product itself, represents a rank-2 tensor. For tensors whose outer product produces a tensor of rank > 2 and type (m, n) where m, n > 0, various contraction operations between different pairs of indices of opposite variance type can occur and hence more than one inner product, which are different in general, can be defined. Moreover, when the outer product produces a tensor of rank > 3 and type (m, n) where m, n > 1, more than one contraction operation can take place simultaneously. Based on what we have seen earlier (refer to § 3.2.3 and 3.2.4), the outer product of a tensor of type (p, q) by a tensor of type (s, t) produces a tensor of type 3.2.6 Permutation 68 (p + s, q + t) and hence (p + s) × (q + t) individual inner product operations can take place, i.e. one inner product for each combination of lower and upper indices. We note that for orthonormal Cartesian systems the variance type is irrelevant and hence a rank-n individual inner product operations which is the number of possible tensor can have n(n−1) 2 contraction operations, as explained earlier (see § 3.2.4). There are more specialized types of inner product; some of these may be defined differently by different authors. For example, a double contraction inner product of two rank-2 tensors, A and B, may be defined and denoted by double vertically- or horizontally-aligned dots (i.e. A : B or A · · B) to indicate double contraction taking place between different pairs of indices. An instance of these specialized types is the inner product with double contraction of two dyads which is commonly defined by: ab : cd = (a · c) (b · d) (138) where the single dots in the right hand side of this equation symbolize the conventional dot product of two vectors. The result of this operation is obviously a scalar since it is the product of two scalars, as can be seen from the right hand side of the equation. Some authors may define a different type of double contraction inner product of two dyads, symbolized by two horizontally-aligned dots, which may be called a transposed contraction. This type of inner product is given by: ab · · cd = ab : dc = (a · d) (b · c) (139) where the result is also a scalar. The second equality of the last equation is based on Eq. 138. We note that the double inner product operators, i.e. : and ··, are defined by some authors opposite to the above definitions (i.e. the other way around) and hence one should be on the lookout for such differences in convention. For two rank-2 tensors, the aforementioned double contraction inner products are similarly defined as in the case of two dyads, that is: A : B = Aij B ij A · · B = Aij B ji (140) Inner products with higher multiplicities of contraction can be defined similarly, and hence they may be regarded as trivial extensions of the inner products with lower contraction multiplicities. Finally, we note that the inner product of tensors produces a tensor because the inner product is an outer product operation followed by a contraction operation and both of these operations on tensors produce tensors, as stated before (see 3.2.3 and 3.2.4). 3.2.6 Permutation A tensor may be obtained by exchanging the indices of another tensor. For example, Aikj is a permutation of the tensor Aijk . A common example of the permutation operation of tensors is the transposition of a matrix representing a rank-2 tensor since the first and second indices, which represent the rows and columns of the matrix, are exchanged in this operation. It is obvious that tensor permutation applies only to tensors of rank > 1 since 3.2.7 Tensor Test and Quotient Rule 69 no exchange of indices can take place in a scalar with no index or in a vector with a single index. Also, permutation may be restricted to indices of the same variance type. The collection of tensors obtained by permuting the indices of a given tensor may be called isomers. 3.2.7 Tensor Test and Quotient Rule Sometimes a tensor-like object may be suspected for being a tensor; in such cases a test based on what is called the “quotient rule” can be used to clarify the situation. We should remark that the quotient rule of tensors must not be confused with the quotient rule of differentiation. According to the quotient rule of tensors, if the inner product of a suspected tensor by a known tensor is a tensor then the suspect is a tensor. In more formal terms, if it is not known if A is a tensor or not but it is known that B and C are tensors; moreover it is known that the following relation holds true in all rotated (i.e. properly-transformed) coordinate frames: Apq...k...m Bij...k...n = Cpq...mij...n (141) then A is a tensor. Here, A, B and C are respectively of ranks m, n and (m+n−2), where the rank of C is reduced by 2 due to the contraction on k which can be any index of A and B independently. We assume, of course, that the rules of contraction of indices, such as being of opposite variance type in the case of non-Cartesian coordinates, are satisfied in this operation. The form given by the above equation is based, for simplicity, on assuming a Cartesian system. Finally, we should remark that testing a suspected tensor for being a tensor can also be done by employing the first principles through direct application of the transformation rules to see if the alleged tensor satisfies the transformation rules of tensors or not. However, using the quotient rule is generally more convenient and requires less work. Another remark is that the quotient rule of tensors may be considered by some authors as a replacement for the division operation which is not defined for tensors. 3.3 Tensor Representations So far, we are familiar with the covariant and contravariant (including mixed) representations of tensors. There is still another type of representation, that is the physical representation which is the common one in the scientific applications of tensor calculus such as fluid and continuum mechanics. The introduction and employment of the physical representation of tensors are justified by the fact that the covariant and contravariant basis vectors, as well as the covariant and contravariant components of a vector, do not in general have the same physical dimensions as explained earlier (see § 2.2). Moreover, the basis vectors may not have the same magnitude. This motivates the introduction of a more standard form of vectors by using physical components (which have the same dimensions) with normalized basis vectors (which are dimensionless with unit magnitude) where the metric tensor and the scale factors are employed to facilitate this process. The 3.3 Tensor Representations 70 normalization of the basis vectors is done by dividing each vector by its magnitude. For example, the normalized covariant and contravariant basis vectors of a general coordinate system, Êi and Êi , are given by:[17] Êi = Ei |Ei | Êi = Ei |Ei | (no sum on i) (142) which for an orthogonal coordinate system becomes (see Eq. 70): Ei Êi = p = hi Ei g ii Ei Ei Êi = √ = gii hi (no sum on i) (143) where gii and g ii are the ith diagonal elements of the covariant and contravariant metric tensor respectively and hi is the scale factor of the ith coordinate as described previously (see § 2.5 and 2.6). Consequently, if the physical components of a contravariant vector are notated with a hat, then for an orthogonal system we have: Ei A = Ai Ei = Âi Êi = Âi √ gii Âi = =⇒ √ gii Ai = hi Ai (no sum) (144) Ai Ai g ii Ai = √ = gii hi (no sum) (145) Similarly, for the covariant form of the vector we have: Ei A = Ai Ei = Âi Êi = Âi p g ii =⇒ Âi = p These definitions and processes can be easily extended to tensors of higher ranks as we will see next. The physical components of higher rank tensors are similarly defined as for rank-1 tensors by considering the basis vectors of the coordinate system of the space where similar simplifications apply to orthogonal coordinate systems. For example, for a rank-2 tensor A in an orthogonal coordinate system, the physical components can be represented by: Âij = Aij hi hj Âij = hi hj Aij hi Aij Âij = hj (no sum on i or j, with basis Êi Êj ) (146) (no sum on i or j, with basis Êi Êj ) (147) (no sum on i or j, with basis Êi Êj ) (148) On generalizing the above pattern, we conclude that the physical components of a tensor of type (m, n) in an orthogonal coordinate system are given by: ...am = Âab11...b n [17] ha1 . . . ham a1 ...am A hb1 . . . hbn b1 ...bn (no sum on any index) (149) We note that the factors in the denominators are scalars and hence the rules of indices are not violated. 3.4 Exercises and Revision 71 where the symbols and basis tensors are defined similarly following the above examples. As a consequence of the last statements, in a space with a well defined metric any tensor can be expressed in covariant or contravariant (including mixed) or physical forms using different sets of basis tensors. Moreover, these forms can be transformed from each other by using the raising and lowering operators and scale factors. As before, for orthonormal Cartesian systems the covariant, contravariant and physical components are the same where the Kronecker delta is the metric tensor. This is because the covariant, contravariant and physical basis vectors are identical in these systems. More generally, for orthogonal coordinate systems the two sets of normalized covariant and contravariant basis vectors are identical as established earlier because the corresponding vectors of each basis set are in the same direction (see § 2.4 and 2.6), and hence the physical components corresponding to the covariant and contravariant components are identical as well. Consequently, for orthogonal coordinate systems with orthonormal basis vectors, the covariant, contravariant and physical components are identical because the normalized bases corresponding to these three forms are identical. The physical components of a tensor may be represented by the symbol of the tensor with subscripts denoting the coordinates of the employed coordinate system. For instance, if A is a vector in a 3D space with contravariant components Ai or covariant components Ai , its physical components in Cartesian, cylindrical, spherical and general curvilinear systems may be denoted by (Ax , Ay , Az ), (Aρ , Aφ , Az ), (Ar , Aθ , Aφ ) and (Au , Av , Aw ) respectively. For consistency and dimensional homogeneity, the tensors in scientific applications are commonly represented by their physical components with a set of normalized basis vectors. The invariance of the tensor form then guarantees that the same tensor formulation is valid regardless of any particular coordinate system where standard tensor transformations can be used to convert from one form to another without affecting the validity and invariance of the formulation. 3.4 Exercises and Revision 3.1 Define “covariant” and “contravariant” tensors from the perspective of their notation and their transformation rules. 3.2 Write the transformation relations for covariant and contravariant vectors and for covariant, contravariant and mixed rank-2 tensors between different coordinate systems. 3.3 State the practical rules for writing the transformation relations of tensors between different coordinate systems. 3.4 What are the raising and lowering operators and how they provide the link between the covariant and contravariant types? 3.5 A is a tensor of type (m, n) and B is a tensor of type (p, q, w). What this means? Write these tensors in their indicial form. 3.6 Write the following equations in full tensor notation and explain their significance: Ei = ∂r ∂ui Ei = ∇ui 3.4 Exercises and Revision 72 3.7 Write the orthonormalized form of the basis vectors in a 2D general coordinate system. Verify that these vectors are actually orthonormal. 3.8 Why the following relations are labeled as the reciprocity relations? Ei · Ej = δij Ei · Ej = δ ij 3.9 The components of the tensors A, B and C are given by: A.ik. j , B.jn. mq and Ckl. i . Write these tensors in their full notation that includes their basis tensors. 3.10 A, B and C are tensors of rank-2, rank-3 and rank-4 respectively in a given coordinate system. Write the components of these tensors with respect to the following basis tensors: , Ei En , Ei Ek Em and Ej Ei Ek En . 3.11 What “dyad” means? Write all the nine unit dyads associated with the double directions of rank-2 tensors in a 3D space with a rectangular Cartesian coordinate system (i.e. e1 e1 · · · e3 e3 ). 3.12 Make a simple sketch of the nine dyads of exercise 3.11. 3.13 Compare true and pseudo vectors making a clear distinction between the two with a simple illustrating plot. Generalize this to tensors of any rank. 3.14 Justify the following statement: “The terms of consistent tensor expressions and equations should be uniform in their true and pseudo type”. 3.15 What is the curl of a pseudo vector from the perspective of true/pseudo qualification? 3.16 Define absolute and relative tensors stating any necessary mathematical relations. 3.17 What is the weight of the product of A and B where A is a tensor of type (1, 2, 2) and B is a tensor of type (0, 3, −1)? 3.18 Show that the determinant of a rank-2 absolute tensor A is a relative scalar and find the weight in the case of A being covariant and in the case of A being contravariant. 3.19 Why the tensor terms of tensor expressions and equalities should have the same weight? 3.20 What “isotropic” and “anisotropic” tensor mean? 3.21 Give an example of an isotropic rank-2 tensor and another example of an anisotropic rank-3 tensor. 3.22 What is the significance of the fact that the zero tensor of all ranks and all dimensions is isotropic with regard to the invariance of tensors under coordinate transformations? 3.23 Define “symmetric” and “anti-symmetric” tensor. Why scalars and vectors are not qualified to be symmetric or anti-symmetric? 3.24 Write the symmetric and anti-symmetric parts of the tensor Aij . 3.25 Write the symmetrization and anti-symmetrization formulae for a rank-n tensor Ai1 ...in . 3.26 Symmetrize and anti- symmetrize the tensor Aijkl with respect to its second and fourth indices. 3.27 Write the two mathematical conditions for a rank-n tensor Ai1 ...in to be totally symmetric and totally anti-symmetric. 3.28 The tensor Aijk is totally symmetric. How many distinct components it has in a 3D space? 3.29 The tensor Bijk is totally anti-symmetric. How many identically vanishing components it has in a 3D space? How many distinct non-identically vanishing components it has 3.4 Exercises and Revision 3.30 3.31 3.32 3.33 3.34 3.35 3.36 3.37 3.38 3.39 3.40 3.41 3.42 3.43 3.44 3.45 3.46 3.47 73 in a 3D space? Give numeric or symbolic examples of a rank-2 symmetric tensor and a rank-2 skewsymmetric tensor in a 4D space. Count the number of independent non-zero components in each case. Write the formula for the number of independent components of a rank-2 symmetric tensor, and the formula for the number of independent non-zero components of a rank-2 anti-symmetric tensor in nD space. Explain why the entries corresponding to identical anti-symmetric indices should vanish identically. Why the indices whose exchange defines the symmetry and anti-symmetry relations should be of the same variance type? Discuss the significance of the fact that the symmetry and anti-symmetry characteristic of a tensor is invariant under coordinate transformations and link this to the invariance of the zero tensor. Verify the relation Aij B ij = 0, where Aij is a symmetric tensor and B ij is an antisymmetric tensor, by writing the sum in full assuming a 3D space. Classify the common tensor operations with respect to the number of tensors involved in these operations. Which of the following operations are commutative, associative or distributive when these properties apply: algebraic addition, algebraic subtraction, multiplication by a scalar, outer multiplication, and inner multiplication? For question 3.37, write all the required mathematical relations that describe those properties. The tensors involved in tensor addition, subtraction or equality should be compatible in their types. Give all the details about these “types”. What is the meaning of multiplying a tensor by a scalar in terms of the components of the tensor? A tensor of type (m1 , n1 , w1 ) is multiplied by another tensor of type (m2 , n2 , w2 ). What is the type, the rank and the weight of the product? We have two tensors: A = Aij Ei Ej and B = B kl Ek El . We also have C = AB and D = BA. Use the properties of tensor operations to obtain the full expression of C and D in terms of their components and basis tensors (i.e. C = AB = · · · etc.). Explain why tensor multiplication, unlike ordinary multiplication of scalars, is not commutative considering the basis tensors to which the tensors are referred. The direct product of vectors a and b is ab. Edit the following equation by adding a simple notation to make it correct without changing the order: ab = ba. What is the difference in notation between matrix multiplication and tensor multiplication of two tensors, A and B, when we write AB? Define the contraction operation of tensors. Why this operation cannot be conducted on scalars and vectors? In reference to general coordinate systems, a single contraction operation is conducted on a tensor of type (m, n, w) where m, n > 0. What is the rank, the type and the weight of the contracted tensor? 3.4 Exercises and Revision 74 3.48 What is the condition that should be satisfied by the two tensor indices involved in a contraction operation assuming a general coordinate system? What about tensors in orthonormal Cartesian systems? 3.49 How many individual contraction operations can take place in a tensor of type (m, n, w) in a general coordinate system? Explain why. 3.50 How many individual contraction operations can take place in a rank-n tensor in an orthonormal Cartesian coordinate system? Explain why. 3.51 List all the possible single contraction operations that can take place in the tensor Aijk lm . 3.52 List all the possible double contraction operations that can take place in the tensor Aij kmn . 3.53 Give examples of contraction operation from matrix algebra. 3.54 Show that contracting a rank-n tensor results in a rank-(n − 2) tensor. 3.55 Discuss inner multiplication of tensors as an operation composed of two more simple operations. 3.56 Give common examples of inner product operation from linear algebra and vector calculus. 3.57 Why inner product operation is not commutative in general? 3.58 Complete the following equations where A and B are rank-2 tensors of opposite variance type: A : B =? A · · B =? 3.59 Write ab : cd in component form assuming a Cartesian system. Repeat this with ab · · cd. 3.60 Why the operation of inner multiplication of tensors results in a tensor? 3.61 We have: A = Ai Ei , B = Bj Ej and C = C mn Em En . Find the following tensor products: AB, AC and BC. 3.62 Referring to the tensors in question 3.61, find the following dot products: B · B, C · A and C · B. 3.63 Define permutation of tensors giving an example of this operation from matrix algebra. 3.64 State the quotient rule of tensors in words and in a formal mathematical form. 3.65 Why the quotient rule is usually used in tensor tests instead of applying the transformation rules? 3.66 Outline the similarities and differences between the three main forms of tensor representation, i.e. covariant, contravariant and physical. 3.67 Define, mathematically, the physical basis vectors, Êi and Êi , in terms of the covariant and contravariant basis vectors, Ei and Ei . hi hj hn ikn 3.68 Correct, if necessary, the following relation: Âikn jm = hk hm Ajm (no sum on any index) where A is a tensor in an orthogonal coordinate system. 3.69 Why the normalized covariant, contravariant and physical basis vectors are identical in orthogonal coordinate systems? 3.70 What is the physical significance of being able to transform one type of tensors to other types as well as transforming between different coordinate systems? 3.71 Why the physical representation of tensors is usually preferred in the scientific appli- 3.4 Exercises and Revision 75 cations of tensor calculus? 3.72 Give a few common examples of physical representation of tensors in mathematical and scientific applications. 3.73 What is the advantage of representing the physical components of a tensor (e.g. A) by the symbol of the tensor with subscripts denoting the coordinates of the employed coordinate system, e.g. (Ar , Aθ , Aφ ) in spherical coordinate systems? Chapter 4 Special Tensors The subject of investigation of this chapter is those tensors that form an essential part of the tensor calculus theory itself, namely the Kronecker, the permutation and the metric tensors. The chapter also includes a section devoted to the generalized Kronecker delta tensor which may be regarded as a bridge between the ordinary Kronecker delta tensor and the permutation tensor. There is also another section in which a number of important tensor identities related to the Kronecker or/and permutation tensors are collected and discussed. Finally, there is a section devoted to some important mathematical definitions and applications in which special tensors are employed. The Kronecker and permutation tensors are of particular importance in tensor calculus due to their distinctive properties and unique transformation attributes. They are numerical tensors with invariant components in all coordinate systems. They enter in the definition of many mathematical objects in tensor calculus and are used to facilitate the formulation of many tensor identities. Similarly, the metric tensor is one of the most important tensors (and may even be the most important) in tensor calculus and its applications, as will be revealed in § 4.5 and other parts of this book. In fact, it permeates the whole subject of tensor calculus due to its role in characterizing the space and its involvement in essential mathematical definitions, operations and transformations. For example, it enters in the definition of many mathematical concepts, such as curve length and surface area, and facilitates the transformation between covariant and contravariant types of tensors and their basis vectors. 4.1 Kronecker delta Tensor As indicated above, there are two types of Kronecker delta tensor: ordinary and generalized. In the present section we discuss the ordinary Kronecker delta tensor and in § 4.4 we investigate the generalized Kronecker delta tensor. The ordinary Kronecker delta tensor, which may also be called the unit tensor, is a rank-2 numeric, absolute, symmetric, constant, isotropic tensor in all dimensions. It is defined in its covariant form as: ( 1 (i = j) δij = (i, j = 1, 2, . . . n) (150) 0 (i 6= j) where n is the space dimension, and hence it can be considered as the identity tensor or matrix. For example, in a 3D space the Kronecker delta tensor is given by:     1 0 0 δ11 δ12 δ13 1 0  (151) [δij ] =  δ21 δ22 δ23  =  0 0 0 1 δ31 δ32 δ33 76 4.2 Permutation Tensor 77 The contravariant and mixed forms of the ordinary Kronecker delta tensor, i.e. δ ij and δij , are similarly defined. Consequently, the numerical values of the components of the covariant, contravariant and mixed types of the Kronecker delta tensor are the same, that is: δij = δ ij = δ ij = δij (152) and hence they are all defined by Eq. 150. We note that in the last equation (and in any similar equation) the numerical values of the components (not the tensors) are equated and hence this is not a violation of the rules of tensor indices as stated in § 1.2. The tensor is made of components with reference to a set of basis vectors and hence the equality of the components of two tensors of different variance type does not imply the equality of the two tensors since the two basis vector sets to which the two tensors are referred can be different. From the above definitions, it can be seen that the Kronecker delta tensor is symmetric, that is: δij = δji δ ij = δ ji (153) where i, j = 1, 2, . . . , n. Moreover, the tensor is conserved under all proper and improper coordinate transformations where “conserved” means that the tensor keeps the numerical values of its components following a coordinate transformation. Since it is conserved under proper transformations, it is an isotropic tensor. We note that being conserved under all transformations is stronger than being isotropic as the former applies even under improper coordinate transformations while isotropy is restricted, by definition, to proper transformations (see § 3.1.4). We also used “conserved” rather than “invariant” to indicate the preservation of the components and to avoid confusion with form-invariance which is a property that characterizes all tensors. 4.2 Permutation Tensor The permutation tensor, which is a numeric tensor with constant components, is also known as the Levi-Civita, anti-symmetric and alternating tensor. We note that the “LeviCivita” label is usually used for the rank-3 permutation tensor. Also some authors distinguish between the permutation tensor and the Levi-Civita tensor even for rank-3. Moreover, some of the common labels and descriptions of the permutation tensor are more specific to rank-3. Hence, differences in conventions, definitions and labels should be considered when reading the literature of tensor calculus related to the permutation tensor. The permutation tensor has a rank equal to the number of dimensions of the space. Hence, a rank-n permutation tensor has nn components. This tensor is characterized by the following properties: 1. It is numeric tensor and hence the value of its components are: −1, 1 and 0 in all coordinate systems. 2. The value of any particular component (e.g. ǫ312 ) of this tensor is the same in any coordinate system and hence it is constant tensor in this sense. 4.2 Permutation Tensor 78 3. It is relative tensor of weight −1 for its covariant form and +1 for its contravariant form. 4. It is isotropic tensor (see § 3.1.4) since its components are conserved under proper transformations. 5. It is totally anti-symmetric in each pair of its indices, i.e. it changes sign on swapping any two of its indices. 6. It is pseudo tensor (see § 3.1.2) since it acquires a minus sign under improper orthogonal transformation of coordinates. 7. The permutation tensor of any rank has only one independent non-vanishing component because all the non-zero components of this tensor are of unity magnitude. 8. The rank-n permutation tensor possesses n! non-zero components which is the number of the non-repetitive permutations of its indices. 9. The rank and the dimension of the permutation tensor are identical and hence in an nD space it has nn components. The rank-2 permutation tensor ǫij in its covariant form is defined by:[18] ǫ12 = 1 ǫ21 = −1 ǫ11 = ǫ22 = 0 Similarly, the rank-3 permutation tensor ǫijk in its covariant form is defined by:   (i, j, k is even permutation of 1,2,3)  1 ǫijk = −1 (i, j, k is odd permutation of 1,2,3)   0 (repeated index) (154) (155) The contravariant form of the rank-2 and rank-3 permutation tensors is similarly defined. Figure 15 is a graphical illustration of the rank-3 permutation tensor ǫijk while Figure 16, which may be used as a mnemonic device, demonstrates the cyclic nature of the three even permutations of the indices of the rank-3 permutation tensor and the three odd permutations of these indices assuming no repetition of indices. The three permutations in each case are obtained by starting from a given number in the cycle and rotating in the given sense to obtain the other two numbers in the permutation. The definition of the rank-n permutation tensor (i.e. ǫi1 i2 ...in and ǫi1 i2 ...in ) is similar to the definition of the rank-3 permutation tensor with regard to the repetition in its indices (i1 , i2 , · · · , in ) and being even or odd permutations in their correspondence to (1, 2, · · · , n), that is:   (i1 , i2 , . . . , in is even permutation of 1, 2, . . . , n)  1 i1 i2 ...in = −1 ǫi1 i2 ...in = ǫ (i1 , i2 , . . . , in is odd permutation of 1, 2, . . . , n) (156)   0 (repeated index) As stated before, equations like this defines the numeric values of the tensor components and hence they do not violate the rules of indices with regard to their variance type (see § 1.2) or the rules of relative tensors with regard to their weight (see § 3.1.3). [18] There is no rank-1 permutation tensor as there is no possibility of permutation in a 1D space. 4.2 Permutation Tensor 79 ǫ332 ǫ331 ǫ321 ǫ322 ǫ311 ǫ232 ǫ231 ǫ221 ǫ222 ǫ211 ǫ131 ǫ121 ǫ111 ǫ132 ǫ122 ǫ112 ǫ212 ǫ233 ǫ312 ǫ333 ǫ323 ǫ313 ǫ223 ǫ213 ǫ133 ǫ123 ǫ113 Figure 15: Graphical illustration of the rank-3 permutation tensor ǫijk where circular nodes represent 0, square nodes represent 1 and triangular nodes represent −1. As well as the inductive definition of the permutation tensor (as given by Eqs. 154, 155 and 156), the permutation tensor of any rank can also be defined analytically where the entries of the tensor are calculated from closed form formulae. Accordingly, the values of the entries of the rank-2 permutation tensor can be calculated from the following closed form equation: ǫij = ǫij = (j − i) (157) where each one of the indices i, j ranges over 1, 2. Similarly, the numerical values of the entries of the rank-3 permutation tensor are given by: ǫijk = ǫijk = 1 (j − i) (k − i) (k − j) 2 (158) where each one of the indices i, j, k ranges over 1, 2, 3. As for the rank-4 permutation 4.2 Permutation Tensor 3 80 3 1 +1 −1 2 2 Even Odd 1 Figure 16: Graphical demonstration of the cyclic nature of the even and odd permutations of the indices of the rank-3 permutation tensor assuming no repetition in indices. tensor we have: ǫijkl = ǫijkl = 1 (j − i) (k − i) (l − i) (k − j) (l − j) (l − k) 12 (159) where each one of the indices i, j, k, l ranges over 1, 2, 3, 4. More generally, the numerical values of the entries of the rank-n permutation tensor can be obtained from the following formula: " # n−1 n Y 1 Y Y 1 ǫa1 a2 ···an = ǫa1 a2 ···an = (aj − ai ) (160) (aj − ai ) = i! S(n − 1) i=1 j=i+1 1≤i<j≤n where each one of the indices a1 , · · · , an ranges over 1, · · · , n and S(n − 1) is the super factorial function of the argument (n − 1) which is defined by: S(k) = k Y i=1 (161) i! = 1! · 2! · . . . · k! A simpler formula for calculating the numerical values of the entries of the rank-n permutation tensor can be obtained from the previous one (Eq. 160) by dropping the magnitude of the multiplication factors and taking their signs only, that is: ! Y Y sgn (aj − ai ) = sgn ǫa1 a2 ···an = ǫa1 a2 ···an = (aj − ai ) (162) 1≤i<j≤n 1≤i<j≤n where sgn(k) is the sign function of the argument k which is defined by:   (k > 0) +1 sgn(k) = −1 (k < 0)   0 (k = 0) (163) 4.2 Permutation Tensor 81 The sign function in Eq. 162 can be expressed in a more direct form for the non-vanishing entries of the permutation tensor (which correspond to the non-repetitive permutations) by dividing each argument of the multiplicative factors in Eq. 162 by its absolute value, noting that none of these factors is zero, and hence Eq. 162 becomes: ǫa1 a2 ···an = ǫa1 a2 ···an = (aj − ai ) |aj − ai | 1≤i<j≤n Y (aj 6= ai ) (164) Regarding the vanishing entries, they are easily identified by having repeated indices and hence all the entries of the tensor are easily evaluated. As stated above, the permutation tensor is totally anti-symmetric (see § 3.1.5) in each pair of its indices, i.e. it changes sign on swapping any two of its indices, that is: ǫi1 ...ik ...il ...in = −ǫi1 ...il ...ik ...in ǫi1 ...ik ...il ...in = −ǫi1 ...il ...ik ...in (165) The reason is that any exchange of two indices requires an even/odd number of singlestep shifts to the right of the first index plus an odd/even number of single-step shifts to the left of the second index, so the total number of shifts is odd and hence it is an odd permutation of the original arrangement. This may also be concluded from the closed form formulae (e.g. Eq. 162) where any exchange will lead to an exchange of the indices of a single multiplicative factor which leads to sign change.[19] Also, the permutation tensor is a pseudo tensor since it acquires a minus sign under an improper orthogonal transformation of coordinates, i.e. inversion of axes that changes the system handedness (see § 2.3.1). However, it is an isotropic tensor since it is conserved under proper coordinate transformations. The permutation tensor may be considered as a contravariant relative tensor of weight +1 or as a covariant relative tensor of weight −1. Since the contravariant and covariant types of the permutation tensor are relative tensors, it is desirable to define absolute covariant and contravariant forms of the permutation tensor. This is done by using the metric tensor, that is: ǫi1 ...in = √ g ǫi1 ...in 1 ǫi1 ...in = √ ǫi1 ...in g (166) where the indexed ǫ and ǫ are respectively the relative and absolute permutation tensors of the given type, and g is the determinant of the covariant metric tensor gij . Referring to Eqs. 63 and 93, we see that the ǫ tensors are absolute with w = 0. We note that the contravariant form of the absolute permutation tensor requires a sign function but the details are out of the scope of the present text (see Zwillinger in the References). However, for the rank-3 permutation tensor, which is the one used mostly in the present book, the above expression stands as it is. [19] The non-existence of a mixed type permutation tensor is also related to its totally anti-symmetric nature. 4.3 Identities Involving Kronecker or/and Permutation Tensors 82 The contravariant and covariant types of the rank-3 permutation tensor are linked, through the Kronecker delta tensor, by the following relations: j i j ǫijk ǫlmk = δli δm − δm δl (167) i δli δm δni j j δl δm δnj k δlk δm δnk (168) ijk ǫ ǫlmn = where the indexed δ represent the mixed form of the ordinary Kronecker delta. The last equation can be generalized to nD spaces as follows: ǫi1 i2 ···in ǫj1 j2 ···jn = δji11 δji12 · · · δji1n δji21 δji22 · · · δji2n . .. . . .. . .. . . δjin1 δjin2 · · · δjinn (169) The contravariant and covariant types of the rank-n permutation tensor are also linked by the following relation: (170) ǫi1 i2 ···in ǫi1 i2 ···in = n! where n is the space dimension. The left hand side of this equation represents the sum of the products of the corresponding entries of the contravariant and covariant types of the permutation tensor. On comparing Eq. 169 and the upcoming Eq. 199 we obtain the following identity which links the contravariant and covariant types of the permutation tensor to each other and to the generalized Kronecker delta: n ǫi1 ...in ǫj1 ...jn = δji11 ...i ...jn (171) More details about these identities will be given in the subsequent sections of this chapter. 4.3 4.3.1 Identities Involving Kronecker or/and Permutation Tensors Identities Involving Kronecker delta Tensor When an index of the Kronecker delta tensor is involved in a contraction operation by repeating an index in another tensor in its own term, the effect of this is to replace the shared index in the other tensor by the other index of the Kronecker delta, that is: δij Aj = Ai δji Aj = Ai (172) In such cases the Kronecker delta is described as an index replacement or substitution operator. Hence, we have: δij δjk = δik (173) 4.3.2 Identities Involving Permutation Tensor Similarly, we have: δij δjk δki = δik δki = δii = n 83 (174) where n is the space dimension. The last part of this equation (i.e. δii = n) can be easily justified by the fact that δii is the trace of the identity tensor considering the summation convention. Due to the fact that the coordinates are independent of each other (see § 2.2), we also have the following identity which is based on the well known rules of partial differentiation: ∂ui = ∂j ui = ui,j = δji ∂uj (175) Hence, in an nD space we obtain the following identity from the identities of Eqs. 175 and 174: ∂i ui = δii = n (176) Based on the above identities and facts, the following identity can be shown to apply in orthonormal Cartesian coordinate systems: ∂xi ∂xj = δij = δji = ∂xj ∂xi (177) This identity is based on the two facts that the coordinates are independent, and the covariant and contravariant types are the same in orthonormal Cartesian coordinate systems. Similarly, for a set of orthonormal vectors, such as the basis set of orthonormal Cartesian system, the following identity can be easily proved: ei · ej = δij (178) where the indexed e represents the orthonormal vectors. This identity is no more than a symbolic statement of the fact that the vectors in such a set are mutually orthogonal and each one is of unit length. We note that for orthonormal basis sets the covariant and contravariant forms are identical as explained earlier. Finally, the double inner product of two dyads formed by an orthonormal set of vectors satisfies the following identity (see § 3.2.5): ei ej : ek el = (ei · ek ) (ej · el ) = δik δjl (179) This identity is a combination of Eq. 138 and Eq. 178. 4.3.2 Identities Involving Permutation Tensor From the definition of the rank-3 permutation tensor, we obtain the following identity which demonstrates the sense of cyclic order of the non-repetitive permutations of this tensor: ǫijk = ǫkij = ǫjki = −ǫikj = −ǫjik = −ǫkji (180) 4.3.2 Identities Involving Permutation Tensor 84 This identity is also a demonstration of the fact that the rank-3 permutation tensor is totally anti-symmetric in all of its indices since a shift of any two indices reverses its sign. Moreover, it reflects the fact that this tensor has only one independent non-zero component since any one of the non-zero entries, all of which are given by Eq. 180, can be obtained from any other one of these entries by at most a reversal of sign. This identity also applies to the contravariant form of the permutation tensor. We also have the following identity for the rank-n permutation tensor: ǫi1 i2 ···in ǫi1 i2 ···in = n! (181) This identity is based on the fact that the left hand side is the sum of the squares of the epsilon symbol over all the n! non-repetitive permutations of n different indices where the value of epsilon of each one of these permutations is either +1 or −1 and hence in both cases their square is +1. The repetitive permutations are zero and hence they do not contribute to the sum. The double inner product of the rank-3 permutation tensor and a symmetric tensor A is given by the following identity: ǫijk Aij = ǫijk Aik = ǫijk Ajk = 0 (182) This is because an exchange of the two indices of A does not affect its value due to the symmetry of A whereas a similar exchange of these indices in ǫijk results in a sign change; hence each term in the sum has its own negative and therefore the total sum is identically zero. This identity also applies to the contravariant permutation tensor with a covariant symmetric tensor. Another identity with a trivial outcome that involves the rank-3 permutation tensor and a vector A is the following: ǫijk Ai Aj = ǫijk Ai Ak = ǫijk Aj Ak = 0 (183) This can be explained similarly by the fact that, due to the commutativity of ordinary multiplication, an exchange of the indices in A’s will not affect the value but a similar exchange in the corresponding indices of ǫijk will cause a change in sign; hence each term in the sum has its own negative and therefore the total sum will be zero. This identity also applies to the contravariant permutation tensor with a covariant vector. We also note that the identities of Eqs. 182 and 183 similarly apply to other ranks of the permutation tensor as they are based on the totally anti-symmetric property of this tensor. Finally, for a set of three orthonormal vectors forming a right handed system, the following identities are satisfied: ei × ej = ǫijk ek ei · (ej × ek ) = ǫijk (184) (185) These identities are based, respectively, on the forthcoming definitions of the cross product (see Eq. 494) and the scalar triple product (see Eq. 495) in tensor notation plus the fact that these vectors are unit vectors. 4.3.3 Identities Involving Kronecker and Permutation Tensors 85 Table 2: Truth table for the identity of Eq. 186. i 1 1 1 1 1 1 1 1 2 2 2 2 2 2 2 2 4.3.3 j 1 1 1 1 2 2 2 2 1 1 1 1 2 2 2 2 k 1 1 2 2 1 1 2 2 1 1 2 2 1 1 2 2 l 1 2 1 2 1 2 1 2 1 2 1 2 1 2 1 2 ǫij 0 0 0 0 1 1 1 1 -1 -1 -1 -1 0 0 0 0 ǫkl 0 1 -1 0 0 1 -1 0 0 1 -1 0 0 1 -1 0 ǫij ǫkl 0 0 0 0 0 1 -1 0 0 -1 1 0 0 0 0 0 δki 1 1 0 0 1 1 0 0 0 0 1 1 0 0 1 1 δlj 1 0 1 0 0 1 0 1 1 0 1 0 0 1 0 1 δli 1 0 1 0 1 0 1 0 0 1 0 1 0 1 0 1 δkj 1 1 0 0 0 0 1 1 1 1 0 0 0 0 1 1 δki δlj − δli δkj 0 0 0 0 0 1 -1 0 0 -1 1 0 0 0 0 0 Identities Involving Kronecker and Permutation Tensors For the rank-2 permutation tensor, we have the following identity which involves the ordinary Kronecker delta tensor in 2D: ǫij ǫkl = δki δli δkj δlj = δki δlj − δli δkj (186) This identity can be proved inductively by building a table for the values on the left and right hand sides as the indices are varied, as seen in Table 2. The pattern of the indices in the determinant array of this identity is simple, that is the indices of the first ǫ provide the indices for the rows as the upper indices of the deltas while the indices of the second ǫ provide the indices for the columns as the lower indices of the deltas. In fact, the role of these indices in indexing the rows and columns can be shifted. This can be explained by the fact that the positions of the two epsilons can be exchanged, since ordinary multiplication is commutative, and hence the role of the epsilons in providing the indices for the rows and columns will be shifted. This can also be done by taking the transposition of the array of the determinant, which does not change the value of the  T determinant since det (A) = det A . We note that a table like Table 2 is similar to the truth tables used in verifying Boolean and logic identities, and for this reason we label it as a “truth table”. We also note that the method of building a “truth table” like Table 2 can also be used for proving other similar identities. The main advantage of this method is that it is easy and straightforward while its main disadvantage is that it is lengthy and hence it may not be feasible for some messy 4.3.3 Identities Involving Kronecker and Permutation Tensors 86 identities. Moreover, it can provide proofs for special cases but it is not general with respect to proving similar identities in a general nD space where n is variable although it may be used as a part of an inductive proof. Another advantage of this method is that it lends itself to programming and hence it is ideal for use in computing. We note that the subject matter in these tables is the tensor components. Another useful identity involving the rank-2 permutation tensor with the Kronecker delta tensor in 2D is the following: ǫil ǫkl = δki (187) This can be obtained from the identity of Eq. 186 by replacing j with l followed by minimal algebraic manipulations using tensor calculus rules, that is: ǫil ǫkl = δki δll − δli δkl (Eq. 186 with j → l) = 2δki − δli δkl = = 2δki δki − (188) (Eq. 174) δki (Eq. 173) Similarly, we have the following identity which correlates the rank-3 permutation tensor to the Kronecker delta tensor in 3D: ijk ǫ ǫlmn = i δli δm δni j j δl δm δnj k δlk δm δnk =  k j k i j k δn + δm δn δl + δni δlj δm δli δm − j k k i j k δl δli δnj δm + δm δl δn + δni δm (189)  Again, the indices in the determinant of this identity follow the same pattern as that of Eq. 186. Another pattern can also be seen in the six terms on the right where the three upper indices of all terms are ijk while the three lower indices in the positive terms are the even permutations of lmn and the three lower indices in the negative terms are the odd permutations of lmn. This identity can also be established by a truth table similar to Table 2. More generally, the determinantal form of Eqs. 186 and 189, which link the rank-2 and rank-3 permutation tensors to the ordinary Kronecker delta tensors in 2D and 3D spaces, can be extended to link the rank-n permutation tensor to the ordinary Kronecker delta tensor in an nD space, that is: ǫi1 i2 ···in ǫj1 j2 ···jn = δji11 δji12 · · · δji1n δji21 δji22 · · · δji2n .. .. . . . . .. . . δjin1 δjin2 · · · δjinn (190) Again, the pattern of the indices in the determinant of this identity in their relation to the indices of the two epsilons follow the same rules as those of Eqs. 186 and 189. Moreover, the proofs of Eqs. 186 and 189 may be extended to Eq. 190 by induction. 4.3.3 Identities Involving Kronecker and Permutation Tensors 87 Another useful identity in this category is the following: ǫijk ǫlmk = i δli δm j j δl δm j i j = δli δm − δm δl (191) This identity can be obtained from the identity of Eq. 189 by replacing n with k, that is: j k k k i j k i j k j k ǫijk ǫlmk = δli δm − δli δkj δm δk + δm δk δl + δki δlj δm − δm δl δk − δki δm δl j j i j i j i j i j i j = 3δl δm + δm δl + δm δl − δl δm − 3δm δl − δl δm j i j = δli δm − δm δl i i δl δm = j δlj δm (192) The pattern of the indices in this identity (Eq. 191) is as before if we exclude the repetitive index k. The identity of Eq. 191, which may be called the epsilon-delta identity or the contracted epsilon identity or the Levi-Civita identity, is very useful in manipulating and simplifying tensor expressions and proving vector and tensor identities; examples of which will be seen in § 7.1.5. We note that the determinantal form, seen in the middle equality of Eq. 191, can be considered as a mnemonic device for this identity where the first and second indices of the first ǫ index the rows while the first and second indices of the second ǫ index the columns, as given above. In fact, the determinantal form in all the above equations which are given in this form is a mnemonic device for all these equations, and not only Eq. 191, where the expanded form, if needed, can be easily obtained from the determinant which can be easily built following the simple pattern of indices, as explained above. Other common identities in this category are: ǫijk ǫljk = 2δli ǫijk ǫijk = 2δii = 2 × 3 = 3! = 6 (193) (194) The first of these identities can be obtained from Eq. 191 with the replacement of m with j followed by some basic tensor manipulation, that is: ǫijk ǫljk = δli δjj − δji δlj = 3δli − δli = (Eq. 191 with m → j) (195) (Eqs. 174 and 173) 2δli while the second of these identities can be obtained from the first by replacing l with i and applying the summation convention in 3D on the right hand side, i.e. using Eq. 174. The second identity is, in fact, an instance of Eq. 181 for a 3D space. Another common identity that involves the rank-3 permutation tensor with the ordinary Kronecker delta in a 3D space is the following: ǫijk δ1i δ2j δ3k = ǫ123 = 1 (196) 4.4 Generalized Kronecker delta Tensor 88 This identity is based in its first part on the use of the ordinary Kronecker delta as an index replacement operator (Eq. 172), where each one of the deltas replaces an index in the permutation tensor, and is based in its second part on the definition of the permutation tensor (Eq. 155). Finally, the following identity can also be obtained from the definition of the rank-3 permutation tensor (Eq. 155) and the use of the ordinary Kronecker delta as an index replacement operator (Eq. 172): ǫijk δji = ǫijk δki = ǫijk δij = ǫijk δkj = ǫijk δik = ǫijk δjk = 0 (197) This is because on replacing one of the indices of the permutation tensor it will have two identical indices and hence it is zero, e.g. ǫijk δji = ǫjjk = 0. We note that identities like Eqs. 196 and 197 also apply to the opposite variance type. Also, Eqs. 196 and 197 apply to the permutation tensor of other ranks with some simple modifications. 4.4 Generalized Kronecker delta Tensor The generalized Kronecker delta tensor in an nD space is an absolute rank-2n tensor of type (n, n) which is normally defined inductively by:   (j1 . . . jn is even permutation of i1 . . . in )  1 i1 ...in δj1 ...jn = −1 (198) (j1 . . . jn is odd permutation of i1 . . . in )   0 (repeated i’s or j’s) It can also be defined analytically by the following n × n determinant: n δji11 ...i ...jn = δji11 δji12 · · · δji1n δji21 δji22 · · · δji2n .. .. . . . . .. . . δjin1 δjin2 · · · δjinn (199) where the δji entries in the determinant are the ordinary Kronecker deltas as defined previously (see § 4.1). In this equation, the pattern of the indices in the generalized n Kronecker delta symbol δji11 ...i ...jn in connection to the indices in the determinant is similar to n the previous patterns seen in § 4.3.3, that is the upper indices in δji11 ...i ...jn provide the upper indices in the ordinary deltas by indexing the rows of the determinant, while the lower n indices in δji11 ...i ...jn provide the lower indices in the ordinary deltas by indexing the columns of the determinant. From the previous identities, it can be shown that: ǫi1 ...in ǫj1 ...jn = δji11 δji12 · · · δji1n δji21 δji22 · · · δji2n .. .. . . . . .. . . δjin1 δjin2 · · · δjinn (200) 4.4 Generalized Kronecker delta Tensor 89 Now, on comparing the last equation with the definition of the generalized Kronecker delta, i.e. Eq. 199, we conclude that: n ǫi1 ...in ǫj1 ...jn = δji11 ...i ...jn (201) Based on Eq. 201, the generalized Kronecker delta is the result of multiplying two relative tensors one of weight w = +1 and the other of weight w = −1 and hence the generalized Kronecker delta has a weight of w = 0. This shows that the generalized Kronecker delta is an absolute tensor, as stated above. We remark that the multiplication of relative tensors produces a tensor whose weight is the sum of the weights of the original tensors (see § 3.1.3). From Eq. 201, we can see that the relation between the rank-n permutation tensor in its covariant and contravariant forms and the generalized Kronecker delta in an nD space is given by: n ǫi1 ...in = ǫ1 ... n ǫi1 ...in = δi11... ...in (202) ...in ǫi1 ...in = ǫi1 ...in ǫ1 ... n = δ i11... n (203) where the first of these equations is obtained from Eq. 201 by substituting (1 . . . n) for (i1 . . . in ) in the two sides with relabeling j as i and noting that ǫ1 ... n = 1, while the second equation is obtained from Eq. 201 by substituting (1 . . . n) for (j1 . . . jn ) in the two sides and noting that ǫ1 ... n = 1. Based on Eqs. 202 and 203, the permutation tensor can be considered as an instance of the generalized Kronecker delta. Consequently, the rank-n permutation tensor can be written as an n × n determinant consisting of the ordinary Kronecker deltas (Eq. 199). Moreover, Eqs. 202 and 203 can provide another definition for the permutation tensor in its covariant and contravariant forms, in addition to the previous inductive and analytic definitions of this tensor which are given by Eqs. 156 and 160. ijk . If we replace n with The 3D generalized Kronecker delta may be symbolized by δlmn k and use the determinantal definition of the generalized Kronecker delta of Eq. 199 followed by conducting a few basic algebraic manipulations using some of the above tensor identities, we obtain: ijk = δlmk i δli δm δki j δlj δm δkj k δlk δm δkk (Eq. 199)    k j k i j k k − δm δl δkj δlk − δlj δkk + δki δlj δm + δm δk − δkj δm = δli δm j k k i j k i j k k j k = δli δm δk − δli δkj δm + δm δk δl − δm δl δk + δki δlj δm − δki δm δl j k j i j i j k i j j = δli δm δk − δli δm + δm δl − δm δl δk + δm δl − δli δm j i j i j j j i j δl − δli δm δl + δm = 3δli δm − δli δm + δm δl − 3δm j = δli δm = δli δlj i j − δm δl i δm j δm (Eq. 173) (Eq. 174) (204) 4.4 Generalized Kronecker delta Tensor 90 ij = δlm (Eq. 199) that is: ij ijk = δlm δlmk (205) Similar identities can be obtained from contracting two corresponding indices of the nD generalized Kronecker delta to obtain (n − 1)D generalized Kronecker delta, e.g. δljij = δli (206) which can be simply verified following the former example, that is: δljij = δli δji δlj δjj = δli δjj − δji δlj = 2δli − δli = δli (207) Returning to the widely used epsilon-delta identity of Eq. 191, if we consider Eq. 205 plus the above identities which correlate the permutation tensor, the generalized Kronecker delta tensor and the ordinary Kronecker delta tensor, then an identity equivalent to Eq. 191 that involves only the generalized and ordinary Kronecker deltas can be obtained, that is:[20] ijk ǫijk ǫlmk = δlmk = = = ij δlm δli δlj j δli δm (Eq. 201) (208) (Eq. 205) i δm j δm (Eq. 199) i j − δm δl This means that the following relation: ijk ij j i j = δlm = δli δm − δm δl δlmk (209) which is no more than the definition of the generalized Kronecker delta of Eq. 199 is another form of the epsilon-delta identity. The pattern of the indices on the right hand side in relation to the indices of the generalized Kronecker delta is very simple, that is we take first the corresponding upper and lower indices followed by the diagonally crossed indices (i.e. | | − ×). It is worth noting that the epsilon-delta identity (Eqs. 191 and 209) can also be expressed in a more general form by employing the metric tensor with the absolute permutation tensor, that is: g ij ǫikl ǫjmn = gkm gln − gkn glm (210) Other identities involving the permutation tensor and the ordinary Kronecker delta tensor can also be formulated in terms of the generalized Kronecker delta tensor. [20] In fact, this is a derivation of Eq. 191. 4.5 Metric Tensor 4.5 91 Metric Tensor The metric tensor, which may also be called the fundamental tensor, is a rank-2 symmetric absolute non-singular tensor, where “non-singular” means invertible and hence its determinant does not vanish at any point in the space. The metric tensor is one of the most important special tensors in tensor calculus, if not the most important of all. Its versatile usage and functionalities permeate the whole discipline of tensor calculus and its applications. One of the main objectives of the metric is to generalize the concept of distance to general coordinate systems and hence maintain the invariance of distance in different coordinate systems, as will be explained next. This tensor is also used to raise and lower indices and thus facilitate the transformation between the covariant and contravariant types. As a tensor, the metric has significance regardless of any coordinate system although it requires a coordinate system to be represented in a specific form (see § 2.7). So, in principle the coordinate system and the space metric are independent entities. In an orthonormal Cartesian coordinate system of an nD space the length of infinitesimal element of arc, ds, connecting two neighboring points in space, one with coordinates xi and the other with coordinates xi + dxi (i = 1, · · · , n), is given by: (ds)2 = dxi dxi = δij dxi dxj (211) This definition of distance is the key to introducing a rank-2 tensor, gij , called the metric tensor which, for a general coordinate system, is defined by: (ds)2 = gij dui duj (212) where the indexed u represents general coordinates. The metric tensor in the last equation is of covariant form. The metric tensor has also a contravariant form which is notated with g ij . It is common to reserve the term “metric tensor” to the covariant form and call the contravariant form, which is its inverse, the associate or conjugate or reciprocal metric tensor. The components of the metric tensor in its covariant and contravariant forms are closely related to the basis vectors, that is: gij = Ei · Ej g ij = Ei · Ej (213) (214) where the indexed E are the covariant and contravariant basis vectors as defined previously (see § 2.6). Because of these relations, the vectors Ei and Ei may be denoted by gi and gi respectively which is more suggestive of their relation to the metric tensor. Similarly, the mixed type metric tensor is given by: g ij = Ei · Ej = δ ij g i j = Ei · Ej = δ i j (215) and hence it is the identity tensor. These equalities, which may be described as the reciprocity relations, represent the fact that the covariant and contravariant basis vectors are reciprocal sets. 4.5 Metric Tensor 92 As a consequence, the covariant metric tensor is given, in full tensor notation, by: gij = ∂xk ∂xk ∂ui ∂uj (216) where xk = xk u1 , . . . , u n  (k = 1, . . . , n) (217) are independent coordinates in an nD space with a rectangular Cartesian system, and ui and uj (i, j = 1, . . . , n) are independent general coordinates. Similarly, for the contravariant metric tensor we have: ∂ui ∂uj g ij = (218) ∂xk ∂xk As stated already, the basis vectors, whether covariant or contravariant, in general coordinate systems are not necessarily mutually orthogonal and hence the metric tensor is not diagonal in general since the dot products given by Eqs. 213 and 214 (or Eqs. 216 and 218) are not necessarily zero when i 6= j. Moreover, since those basis vectors vary in general in their magnitude and relative orientations and they are not necessarily of unit length, the entries of the metric tensor, including the diagonal elements, are not necessarily of unit magnitude. Also, the entries of the metric tensor, including the diagonal elements, can be positive or negative.[21] However, since the dot product of vectors is a commutative operation, the metric tensor is necessarily symmetric. We note that the mixed type metric tensor is diagonal (or in fact the unity tensor) because the covariant and contravariant basis vector sets are reciprocal systems (see Eq. 215). This applies in all coordinate systems. As indicated above, the covariant and contravariant forms of the metric tensor are inverses of each other and hence we have the following relations:  ij   −1 (219) g = [gij ]−1 [gij ] = g ij Hence: g ik gkj = δ ij gik g kj = δi j (220) where these equations can be seen to represent matrix multiplication (row×column). A result that can be obtained from the previous statements plus Eqs. 138, 213 and 214 is that:  (221) Ei Ej : Ej Ek = Ei · Ej (Ej · Ek ) = g ij gjk = δ ik  jk k j k j k = gij g = δi (222) Ei E : Ej E = (Ei · Ej ) E · E Since the metric tensor has an inverse, it should be non-singular and hence its determinant, which in general is a function of coordinates like the metric tensor itself, should not vanish at any point in the space, that is: g(u1 , . . . , un ) = det (gij ) 6= 0 [21] (223) The diagonal entries can be negative when the coordinates are imaginary (see § 2.2.3 and Eq. 241). 4.5 Metric Tensor 93 From the previous statements, it may be concluded that the metric tensor is in fact a transformation of the Kronecker delta tensor in its different variance types from an orthonormal Cartesian coordinate system to a general coordinate system, that is (see Eqs. 73-75): ∂xk ∂xl δkl = ∂ui ∂uj ∂ui ∂uj kl δ = g ij = ∂xk ∂xl ∂ui ∂xl k gji = δ = ∂xk ∂uj l gij = ∂xk ∂xk = Ei · Ej ∂ui ∂uj ∂ui ∂uj = Ei · Ej k k ∂x ∂x ∂ui ∂xk = Ei · Ej ∂xk ∂uj (covariant) (224) (contravariant) (225) (mixed) (226) As stated above, the metric tensor is symmetric in its two indices, that is: g ij = g ji gij = gji (227) This can be easily explained by the commutativity of the dot product of vectors in reference to the above equations involving the dot product of the basis vectors (Eqs. 213 and 214). Because of the relations: Ai = A · Ei = Aj Ej · Ei = Aj g ji Ai = A · Ei = Aj Ej · Ei = Aj gji (228) (229) the metric tensor is used as an operator for raising and lowering indices and hence facilitating the transformation between the covariant and contravariant types of vectors. By a similar argument, the above can be generalized where the contravariant metric tensor is used for raising covariant indices of covariant and mixed tensors and the covariant metric tensor is used for lowering contravariant indices of contravariant and mixed tensors of any rank, e.g. Aik = g ij Ajk Aikl = gij Ajkl (230) Consequently, any tensor in a Riemannian space with well-defined metric can be cast into covariant or contravariant or mixed forms where for the mixed form the rank should be > 1. We note that in the operations of raising and lowering of indices the metric tensor acts, like the Kronecker delta tensor, as an index replacement operator beside its action in shifting the index position. In this context, it should be emphasized that the order of the raised and lowered indices is important and hence: g ik Ajk = Aj i and g ik Akj = Ai j (231) are different in general. A dot may be used to indicate the original position of the shifted index and hence the order of the indices is recorded, e.g. Aj i· and Ai· j for the above examples respectively. Because raising and lowering of indices is a reversible process, keeping a record of the original position of the shifted indices will facilitate the reversal if 4.5 Metric Tensor 94 needed. We note that dots may also be inserted in the symbols of mixed tensors to remove any ambiguity about the order of the indices even without the action of the raising and lowering operators (refer to § 1.2). For a space with a coordinate system in which the metric tensor can be cast into a diagonal form with all the diagonal entries being of unity magnitude (i.e. ±1) the space and the metric are called flat. For example, in 2D manifolds a plane surface is a flat space since it can be coordinated by an orthonormal 2D Cartesian system which results into a diagonal unity metric tensor since the basis vectors of this system are mutually perpendicular constant vectors and each is of unit length. On the other hand, an ellipsoid is not a flat space (i.e. it is curved) because due to its intrinsic curvature it is not possible to find a coordinate system for this type of surface whose basis vectors produce a diagonal metric tensor with constant diagonal elements of unity magnitude. In this context we note that a Riemannian metric, gij , in a particular coordinate system is a Euclidean metric if it can be transformed to the identity tensor, δij , by a permissible coordinate transformation. If g and ḡ are the determinants of the covariant metric tensor in unbarred and barred coordinate systems respectively, i.e. g = det (gij ) and ḡ = det (ḡij ), then we have: √ √ ḡ = J 2 g ḡ = J g (232) where J (= ∂u ) is the Jacobian of the transformation between the unbarred and barred ∂ ū systems. Consequently, the determinant of the covariant metric tensor and its square root are relative scalars of weight +2 and +1 respectively (see § 3.1.3). A “conjugate” or “associate” tensor of a tensor in a metric space is a tensor obtained by inner product multiplication, once or more, of the original tensor by the covariant or contravariant forms of the metric tensor. All tensors associated with a particular tensor through the metric tensor represent the same tensor but in different base systems since the association is no more than raising or lowering indices by the metric tensor which is equivalent to a representation of the components of the tensor relative to different basis sets. A sufficient and necessary condition for the components of the metric tensor to be constants in a given coordinate system is that the Christoffel symbols of the first or second kind vanish identically (refer to 5.1). This may be concluded from Eqs. 307 and 308. The metric tensor behaves as a constant with respect to covariant and absolute differentiation (see § 5.2 and § 5.3). Hence, in all coordinate systems the covariant and absolute derivatives of the metric tensor are zero. Accordingly, the covariant and absolute derivative operators bypass the metric tensor in differentiating inner and outer products of tensors involving the metric tensor. In orthogonal coordinate systems in an nD space the metric tensor in its covariant and contravariant forms is diagonal with non-vanishing diagonal elements, that is: gij = 0 gii 6= 0 g ij = 0 ii g 6= 0 (i 6= j) (no sum on i) (233) (234) 4.5 Metric Tensor 95 Moreover, we have: gii = (hi )2 = 1 g ii det (gij ) = g = g11 g22 . . . gnn = (no sum on i) Y (hi )2 (235) (236) i det g  ij " Y 1 1 = (hi )2 = = g g11 g22 . . . gnn i #−1 (237) where hi (= |Ei |) are the scale factors, as described previously in § 2.5 and 2.6. As indicated before, for orthonormal Cartesian coordinate systems in a 3D space, the metric tensor is given in its covariant and contravariant forms by the 3 × 3 unit matrix, that is:   1 0 0     1 0  = δ ij = g ij (238) [gij ] = [δij ] =  0 0 0 1 For cylindrical coordinate systems of 3D spaces identified by the coordinates (ρ, φ, z), the metric tensor is given in its covariant and contravariant forms by:     1 0 0 1 0 0   1 0  ρ2 0  g ij =  0 (239) [gij ] =  0 ρ2 0 0 1 0 0 1 while for spherical coordinate systems of 3D spaces identified by the coordinates (r, θ, φ), the metric tensor is given in its covariant and contravariant forms by:     1 0 0 1 0 0  ij    0 0 g =  0 r12 [gij ] =  0 r2 (240) 2 1 2 0 0 r2 sin2 θ 0 0 r sin θ As seen in Eqs. 238-240, all these metric tensors are diagonal since all these coordinate systems are orthogonal. We also notice that all the corresponding diagonal elements of the covariant and contravariant types are reciprocals of each other. This can be easily explained by the fact that these two types are inverses of each other, moreover the inverse of an invertible diagonal matrix is a diagonal matrix obtained by taking the reciprocal of the corresponding diagonal elements of the original matrix, as it is known from linear algebra. We also see that the diagonal elements are the squares of the scale factors of these systems or their reciprocals (refer to Table 1). The Minkowski metric, which is the metric tensor of the 4D space-time of the mechanics of Lorentz transformations, is given by one of the following two forms:     1 0 0 0 −1 0 0 0    0 −1  ij   0 0 0  1 0 0     (241) = [gij ] = g ij =  [g ] = g ij  0  0 0 −1 0  0 1 0  0 0 0 −1 0 0 0 1 4.5 Metric Tensor 96 Consequently, the length of line element ds can be imaginary (see Eqs. 36 and 212). As seen, this metric belongs to a flat space since it is diagonal with all the diagonal entries being ±1. The partial derivatives of the components of the covariant and contravariant metric tensor satisfy the following identities: ∂k gij = −gmj gni ∂k g nm ∂k g ij = −g mj g in ∂k gnm (242) (243) A related formula for the partial derivatives of the components of the covariant and contravariant metric tensor is given by: gim ∂k g mj = −g mj ∂k gim (244) This relation can be obtained by partial differentiation of the relation gim g mj = δij (Eq. 220) with respect to the k th coordinate using the product rule, that is:  =⇒ gim ∂k g mj = −g mj ∂k gim (245) ∂k gim g mj = gim ∂k g mj + g mj ∂k gim = ∂k δij = 0 The last step in the differentiation (i.e. ∂k δij = 0) is justified by the fact that the components of the Kronecker delta tensor are constants. In fact, Eq. 242 can be obtained form Eq. 244 by relabeling m as n and multiplying both sides of Eq. 244 with gmj followed by contraction and exchanging the labels of j and m, that is: g mj ∂k gim = −gim ∂k g mj nj g ∂k gin = −gin ∂k g nj nj gmj g ∂k gin = −gmj gin ∂k g nj n δm ∂k gin = −gmj gin ∂k g nj nj ∂k gim = −gmj gin ∂k g ∂k gij = −gmj gni ∂k g nm (Eq. 244) (246) (m → n) (247) (×gmj ) (248) (Eq. 220) (249) (Eq. 172) (m ↔ j) (250) (251) Similarly, Eq. 243 can be obtained form Eq. 244 by multiplying both sides of Eq. 244 with g in followed by contraction and exchanging the labels of i and n, that is: in gim ∂k g mj = −g mj ∂k gim g gim ∂k g mj in mj = −g g ∂k gim n δm ∂k g mj = −g in g mj ∂k gim ∂k g nj ij in mj = −g g ∂k g = −g mj in ∂k gim g ∂k gnm (Eq. 244) (252) in (×g ) (253) (Eq. 220) (254) (Eq. 172) (255) (n ↔ i) (256) 4.6 Definitions Involving Special Tensors 4.6 97 Definitions Involving Special Tensors In the following subsections, we investigate a number of mathematical objects and operations whose definitions and applications are dependent on the above described special tensors, particularly the permutation and metric tensors. These are just a few examples of tensor definitions involving the special tensors and hence they are not comprehensive in any way. 4.6.1 Dot Product The dot product of two basis vectors in general coordinate systems was given earlier in § 4.5 (see Eqs. 213-215). This will be used in the present subsection to develop expressions for the dot product of vectors and tensors in general. The dot product of two vectors, A and B, in general coordinate systems using their covariant and contravariant forms, as well as opposite forms, is given by: A · B = Ai Ei · Bj Ej = Ai Bj Ei · Ej = g ij Ai Bj = Aj Bj = Ai B i A · B = Ai Ei · B j Ej = Ai B j Ei · Ej = gij Ai B j = Aj B j = Ai Bi A · B = Ai Ei · B j Ej = Ai B j Ei · Ej = δ ij Ai B j = Aj B j A · B = Ai Ei · Bj Ej = Ai Bj Ei · Ej = δij Ai Bj = Ai Bi (257) (258) (259) (260) In brief, the dot product of two vectors is the dot product of their two basis vectors multiplied algebraically by the algebraic product of their components. Because the dot product of the basis vectors is a metric tensor, the metric tensor will act on the components by raising or lowering the index of one component or by replacing the index of a component, as seen in the above equations. The dot product operations outlined in the previous paragraph can be easily extended to tensors of higher ranks where the covariant and contravariant forms of the components and basis vectors are treated in a similar manner to the above examples to obtain the dot product. For instance, the dot product of a rank-2 tensor of contravariant components Aij and a vector of covariant components Bk is given by:    (261) A · B = Aij Ei Ej · Bk Ek = Aij Bk Ei Ej · Ek = Aij Bk Ei δjk = Aij Bj Ei that is, the ith component of this product, which is a contravariant vector, is: [A · B]i = Aij Bj (262) From the previous statements, we conclude that the dot product in general coordinate systems occurs between two vectors of opposite variance type. Therefore, to obtain the dot product of two vectors of the same variance type, one of the vectors should be converted to the opposite type by the raising or lowering operator, followed by the inner product operation. This can be generalized to the dot product of higher-rank tensors where the two contracted indices of the dot product should be of opposite variance type and hence 4.6.2 Magnitude of Vector 98 the index shifting operator in the form of the metric tensor should be used, if necessary, to achieve this. We note that the generalized dot product of two tensors is invariant under permissible coordinate transformations. We also note that the variance type of the tensors involved in an inner product operation is irrelevant for orthonormal Cartesian systems, as explained before. 4.6.2 Magnitude of Vector The magnitude of a contravariant vector A is given by: q q p √ p p |A| = A · A = (Ai Ei ) · (Aj Ej ) = (Ei · Ej ) Ai Aj = gij Ai Aj = Aj Aj = Ai Ai (263) where Eqs. 213 and 229 are used. A similar expression can be obtained for the covariant form of the vector, that is: q q p √ p p i j |A| = A · A = (Ai E ) · (Aj E ) = (Ei · Ej ) Ai Aj = g ij Ai Aj = Aj Aj = Ai Ai (264) where Eqs. 214 and 228 are used. The magnitude of a vector can also be obtained more directly from the dot product of the covariant and contravariant forms of the vector, that is: q q q p √ p |A| = A · A = (Ai Ei ) · (Aj Ej ) = (Ei · Ej ) Ai Aj = δji Ai Aj = Ai Ai = Aj Aj (265) where Eqs. 215 and 172 are used. 4.6.3 Angle between Vectors The angle θ between two contravariant vectors A and B is given by: A·B Ai Ei · B j Ej gij Ai B j =√ k =p cos θ = √ √ |A| |B| A E k · Al E l B m E m · B n E n gkl Ak Al gmn B m B n Similarly, the angle θ between two covariant vectors A and B is given by: A·B Ai Ei · Bj Ej g ij Ai Bj p =√ = √ √ |A| |B| Ak Ek · Al El Bm Em · Bn En g kl Ak Al g mn Bm Bn For two vectors of opposite variance type we have: cos θ = A·B Ai B i Ai B i =p =p √ √ |A| |B| gkl Ak Al g mn Bm Bn g kl Ak Al gmn B m B n All these three cases can be represented by the following formula: cos θ = (266) (267) (268) Aj B j Ai Bi √ = (269) √ √ Al Al Bn B n Al Al Bn B n The angle θ between two sufficiently smooth space curves, C1 and C2 , intersecting at a given point P in the space is defined as the angle between their tangent vectors, A and B, at that point (see Figure 17). cos θ = √ 4.6.4 Cross Product 99 B θ C1 A P C2 Figure 17: The angle between two space curves, C1 and C2 , as the angle θ between their tangents, A and B, at the point of intersection P . 4.6.4 Cross Product The cross product of two covariant basis vectors in general coordinate systems of a 3D space is given by: Ei × Ej = ∂xm ∂xl ∂xm ∂xl ∂xm ∂xl e × e = e × e = ǫlmn en l m l m ∂ui ∂uj ∂ui ∂uj ∂ui ∂uj (270) where the indexed x and u are the coordinates of Cartesian and general coordinate systems respectively, the indexed e are the Cartesian basis vectors and ǫlmn is the rank-3 permutation relative tensor as defined by Eq. 155. In the last step of the last equation, Eq. 184 is used to express the cross product of two orthonormal vectors in tensor notation. We note that for orthonormal Cartesian systems, there is no difference between covariant and contravariant tensors and hence ei = ei . We also note that for orthonormal Cartesian systems g = 1 where g is the determinant of the covariant metric tensor (see Eqs. 238 and 236). n Now since en = en = ∂x Ek , the last equation becomes: ∂uk Ei × Ej = ∂xl ∂xm ∂xn ǫlmn Ek = ǫijk Ek ∂ui ∂uj ∂uk (271) where the underlined absolute covariant permutation tensor is defined as: ǫijk = ∂xl ∂xm ∂xn ǫlmn ∂ui ∂uj ∂uk (272) So the final result is: Ei × Ej = ǫijk Ek (273) 4.6.5 Scalar Triple Product 100 By a similar reasoning, we obtain the following expression for the cross product of two contravariant basis vectors in general coordinate systems: Ei × Ej = ǫijk Ek (274) where the absolute contravariant permutation tensor is defined as: ǫijk = ∂ui ∂uj ∂uk lmn ǫ ∂xl ∂xm ∂xn (275) Considering Eq. 166, the above equations can also be expressed as: Ei × Ej = ǫijk Ek = √ gǫijk Ek (276) ijk ǫ Ei × Ej = ǫijk Ek = √ Ek g (277) The cross product of non-basis vectors follows similar rules to those outlined above for the basis vectors; the only difference is that the algebraic product of the components is used as a scale factor for the cross product of their basis vectors. For example, the cross product of two contravariant vectors, Ai and B j , is given by:   (278) A × B = Ai Ei × B j Ej = Ai B j (Ei × Ej ) = ǫijk Ai B j Ek that is, the k th component of this product, which is a vector with covariant components, is: (279) [A × B]k = ǫijk Ai B j Similarly, the cross product of two covariant vectors, Ai and Bj , is given by:    A × B = Ai Ei × Bj Ej = Ai Bj Ei × Ej = ǫijk Ai Bj Ek (280) with the k th contravariant component being given by: [A × B]k = ǫijk Ai Bj 4.6.5 (281) Scalar Triple Product The scalar triple product of three contravariant vectors in a 3D space is given by:  (282) A · (B × C) = Ai Ei · B j Ej × C k Ek = Ai B j C k Ei · (Ej × Ek )  = Ai B j C k Ei · ǫjkl El  = ǫjkl Ai B j C k Ei · El (Eq. 273) = ǫjkl Ai B j C k δil (Eq. 215) = ǫjki Ai B j C k (Eq. 172) 4.6.6 Vector Triple Product 101 = ǫijk Ai B j C k (Eq. 180) Similarly, the scalar triple product of three covariant vectors in a 3D space is given by:  (283) A · (B × C) = Ai Ei · Bj Ej × Ck Ek  = Ai Bj Ck Ei · Ej × Ek  (Eq. 274) = Ai Bj Ck Ei · ǫjkl El  i jkl = ǫ Ai B j C k E · E l = ǫjkl Ai Bj Ck δli (Eq. 215) jki (Eq. 172) ijk (Eq. 180) = ǫ Ai B j C k = ǫ Ai B j C k Other forms can also be obtained by a similar method. 4.6.6 Vector Triple Product The vector triple product of three vectors, one contravariant and two covariant, in a 3D space is given by:  (284) A × (B × C) = Ai Ei × Bj Ej × Ck Ek  = Ai B j C k E i × E j × E k  (Eq. 274) = Ai Bj Ck Ei × ǫjkl El = ǫjkl Ai Bj Ck (Ei × El ) = ǫjkl Ai Bj Ck ǫilm Em (Eq. 273) = ǫilm ǫjkl Ai Bj Ck Em = ǫilm ǫjkl Ai Bj Ck Em (Eq. 166) Following relabeling of indices and writing in a covariant component form, we obtain the following more organized expression: [A × (B × C)]i = ǫijk ǫklm Aj Bl Cm (285) We can also obtain a different form from one covariant vector and two contravariant vectors, that is:  (286) A × (B × C) = Ai Ei × B j Ej × C k Ek = Ai B j C k Ei × (Ej × Ek )  = Ai B j C k Ei × ǫjkl El  = ǫjkl Ai B j C k Ei × El = ǫjkl Ai B j C k ǫilm Em (Eq. 273) (Eq. 274) 4.6.7 Determinant of Matrix 102 = ǫilm ǫjkl Ai B j C k Em = ǫilm ǫjkl Ai B j C k Em (Eq. 166) Following relabeling of indices and writing in a contravariant component form, we obtain the following more organized expression: [A × (B × C)]i = ǫijk ǫklm Aj B l C m (287) The expressions of the other principal form of the vector triple product, i.e. (A × B) × C, can be obtained by a similar method. 4.6.7 Determinant of Matrix For a 3 × 3 matrix representing a rank-2 tensor A of mixed form in a 3D space, the determinant is given by: det (A) = A11 A12 A13 A21 A22 A23 A31 A32 A33 = 1 ijk 1 ijk l m n n ǫ ǫlmn Ali Am δ AA A j Ak = 3! 3! lmn i j k (288) where Eq. 201 is used in the last step. This may be expressed as a row expansion by substituting 1, 2, 3 for l, m, n that is: det (A) = ǫijk A1i A2j A3k (289) It can also be expressed as a column expansion by substituting 1, 2, 3 for i, j, k that is: n det (A) = ǫlmn Al1 Am 2 A3 (290) More generally, for an n × n matrix representing a rank-2 tensor A of mixed form in an nD space, the determinant is given by: 1 i1 ···in ǫj1 ···jn Aji11 . . . Ajinn (291) ǫ n! Similar definitions can be given using the covariant and contravariant forms of the tensor with the employment of the opposite variance type of the permutation tensors. det (A) = 4.6.8 Length The differential of displacement vector in general coordinate systems is given by: X Ei i X ∂r i i du = E du = |E | du = |Ei | Êi dui dr = i i i ∂u |Ei | i i (292) where r is the position vector as defined previously and the hat indicates a normalized vector. The length of line element, ds, which may also be called the differential of arc length, in general coordinate systems is given by: (ds)2 = dr · dr = Ei dui · Ej duj = (Ei · Ej ) dui duj = gij dui duj (293) 4.6.9 Area 103 where gij is the covariant metric tensor. For orthogonal coordinate systems, the metric tensor is given by: ( 0 (i 6= j) gij = 2 (hi ) (i = j) (294) where hi is the scale factor of the ith coordinate. Hence, the last part of Eq. 293 becomes: X (ds)2 = (hi )2 dui dui (295) i with no cross terms (i.e. terms of products involving more than one coordinate like dui duj where i 6= j) which are generally present in the case of non-orthogonal coordinate systems. On conducting a transformation from one coordinate system to another coordinate system, where the other system is marked with barred coordinates ū, the length of line element ds will be expressed in the new system as: (ds)2 = ḡij dūi dūj (296) Since the length of line element is an invariant quantity, the same symbol (ds)2 is used in both Eqs. 293 and 296. Based on the above formulations, the length L of a t-parameterized space curve C(t) defined by ui = ui (t) where i = 1, · · · , n, which represents the distance traversed along the curve on moving between its start point P1 and end point P2 , is given in general coordinate systems of an nD space by: ˆ t2 r ˆ P2 p ˆ dui duj i j dt (297) gij du du = gij ds = L= dt dt P1 C t1 where C represents the space curve, t is a scalar real variable, and t1 and t2 are the values of t corresponding to the start and end points respectively. It is noteworthy that some authors add a sign indicator to ensure that the argument of the square root in the above equation is positive. However, as indicated in § 1.1, such a condition is assumed when needed since in this book we deal with real quantities only. 4.6.9 Area In general coordinate systems of a 3D space, the area of an infinitesimal element of the surface u1 = c1 in the neighborhood of a given point P , where c1 is a constant, is obtained by taking the magnitude of the cross product of the differentials of the displacement vectors in the directions of the other two coordinates on that surface at P , i.e. the tangent vectors to the other two coordinate curves at P (see Figure 18). Hence, the area of a differential element on the surface u1 = c1 is given by: dσ(u1 = c1 ) = |dr2 × dr3 | (298) 4.6.9 Area 104 ∂r ∂r × 3 du2 du3 2 ∂u ∂u = |E2 × E3 | du2 du3 = 1 2 = ǫ231 E du du (Eq. 45) 3 (Eq. 273) = |ǫ231 | E1 du2 du3 √ √ = g E1 · E1 du2 du3 √ p = g g 11 du2 du3 p = gg 11 du2 du3 (Eqs. 166 & 263) (Eq. 214) where σ represents area and dr2 and dr3 are the displacement differentials along the second and third coordinate curves at P while the other symbols are as defined previously. dr2 × dr3 dr3 dσ u2 P dr2 u3 Figure 18: The area dσ of an infinitesimal element of the surface u1 = c1 in the neighborhood of a given point P as the magnitude of the cross product of the differentials of the displacement vectors in the directions of the other two coordinate curves on that surface at P , dr2 and dr3 . On generalizing the above argument, the area of a differential element on the surface ui = ci (i = 1, 2, 3) in a 3D space where ci is a constant is given by: p (i 6= j 6= k, no sum on i) (299) dσ(ui = ci ) = gg ii duj duk In orthogonal coordinate systems in a 3D space we have: s p 1 = hj hk (i 6= j 6= k, no sum on any index) (300) gg ii = (hi )2 (hj )2 (hk )2 (hi )2 where Eqs. 235 and 236 are used. Hence, Eq. 299 becomes: dσ(ui = ci ) = hj hk duj duk (i 6= j 6= k, no sum on any index) (301) 4.6.10 Volume 105 The last formula represents the area of a surface element of a rectangular shape with sides hj duj and hk duk (no sum on j or k). Based on the above discussion, the area A of a finite surface patch can be defined by the following formula: ¨ A= dσ (302) S where S represents a surface patch and dσ is the area differential of an infinitesimal element on the patch. For coordinate surfaces, the expression for dσ can be obtained from the previous formulations. 4.6.10 Volume In general coordinate systems of a 3D space, the volume of an infinitesimal element of a solid body occupying a given region of the space in the neighborhood of a given point P , where the element is represented by a parallelepiped defined by the three differentials of the displacement vectors dri (i = 1, 2, 3) along the three coordinate curves at P , is obtained by taking the magnitude of the scalar triple product of these vectors (refer to § 4.6.5 and see Figure 19). Hence, the volume of a differential element of the body is given by: dτ = |dr1 · (dr2 × dr3 )|   ∂r ∂r ∂r · × 3 du1 du2 du3 = 1 2 ∂u ∂u ∂u = |E1 · (E2 × E3 )| du1 du2 du3 1 1 2 = E1 · ǫ231 E du du du 3 = E1 · E1 |ǫ231 | du1 du2 du3 = δ11 |ǫ231 | du1 du2 du3 √ = g du1 du2 du3 = J du1 du2 du3 (303) (Eq. 45) (Eq. 273) (Eq. 215) (Eq. 166) (Eq. 63) where g is the determinant of the covariant metric tensor gij , and J is the Jacobian of the transformation as defined previously (see § 2.3). We note that due to the freedom of choice of the order of the variables, which is related to the choice of the coordinate system handedness that could affect the sign of the determinant Jacobian, the sign of the determinant should be adjusted if necessary to have a proper sign for the volume element as a positive quantity. The last line in the last equation is particularly useful for changing the variables in multivariate integrals where the Jacobian facilitates the transformation. In orthogonal coordinate systems in a 3D space, the above formulation becomes: dτ = h1 h2 h3 du1 du2 du3 (304) where h1 , h2 and h3 are the scale factors and where Eq. 236 is used. Geometrically, the last formula represents the volume of a rectangular parallelepiped with edges h1 du1 , h2 du2 and h3 du3 . 4.6.10 Volume 106 dr3 dr2 dr1 u1 P u2 u3 Figure 19: The volume of an infinitesimal element of a solid body in a 3D space in the neighborhood of a given point P as the magnitude of the scalar triple product of the differentials of the displacement vectors in the directions of the three coordinate curves at P , dr1 , dr2 and dr3 . The formulae of Eq. 303, which are specific to a 3D space, can be extended to the differential of a generalized volume element in general coordinate systems of an nD space as follows: √ (305) dτ = gdu1 . . . dun = J du1 . . . dun We note that generalized volume elements are used, for instance, to represent the change of variables in multi-variable integrations. Based on the above discussion, the volume V of a solid body occupying a finite region of the space can be defined as: ˚ V = dτ (306) Ω where Ω represents the region of the space occupied by the solid body and dτ is the volume differential of an infinitesimal element of the body, and where the expression for dτ should be obtained from the above formulations. 4.7 Exercises and Revision 4.7 107 Exercises and Revision 4.1 What is special about the Kronecker delta, the permutation and the metric tensors and why they deserve special attention? 4.2 Give detailed definition, in words and in symbols, of the ordinary Kronecker delta tensor in an nD space. 4.3 List and discuss all the main characteristics (e.g. symmetry) of the ordinary Kronecker delta tensor. 4.4 Write the matrix that represents the ordinary Kronecker delta tensor in a 4D space. 4.5 Do we violate the rules of tensor indices when we write: δij = δ ij = δ ij = δij ? 4.6 Explain the following statement: “The ordinary Kronecker delta tensor is conserved under all proper and improper coordinate transformations”. What is the relation between this and the property of isotropy of this tensor? 4.7 List and discuss all the main characteristics (e.g. anti-symmetry) of the permutation tensor. 4.8 What are the other names used to label the permutation tensor? 4.9 Why the rank and the dimension of the permutation tensor are the same? Accordingly, what is the number of components of the rank-2, rank-3 and rank-4 permutation tensors? 4.10 Why the permutation tensor of any rank has only one independent non-vanishing component? 4.11 Prove that the rank-n permutation tensor possesses n! non-zero components. 4.12 Why the permutation tensor is totally anti-symmetric? 4.13 Give the inductive mathematical definition of the components of the permutation tensor of rank-n. 4.14 State the most simple analytical mathematical definition of the components of the permutation tensor of rank-n. 4.15 Make a sketch of the array representing the rank-3 permutation tensor where the nodes of the array are marked with the symbols and values of the components of this tensor. 4.16 Define, mathematically, the rank-n covariant and contravariant absolute permutation tensors, ǫi1 ...in and ǫi1 ...in . 4.17 Show that ǫijk is a relative tensor of weight −1 and ǫijk is a relative tensor of weight +1. √ 4.18 Show that ǫi1 ...in = g ǫi1 ...in and ǫi1 ...in = √1g ǫi1 ...in are absolute tensors assuming that ǫi1 ...in is a relative tensor of weight −1 and ǫi1 ...in is a relative tensor of weight +1. 4.19 Write ǫi1 i2 ···in ǫj1 j2 ···jn in its determinantal form in terms of the ordinary Kronecker delta. 4.20 Prove the following identity: ǫi1 i2 ···in ǫi1 i2 ···in = n!. 4.21 State a mathematical relation representing the use of the ordinary Kronecker delta tensor acting as an index replacement operator. 4.22 Prove the following relation inductively by writing it in an expanded form in a 3D space: δii = n. 4.7 Exercises and Revision 108 4.23 Repeat exercise 4.22 with the relation: ui,j = δji using a matrix form. 4.24 Justify the following relation assuming an orthonormal Cartesian system: ∂i xj = ∂j xi . 4.25 Justify the following relations where the indexed e are orthonormal vectors: ei · ej = δij ei ej : ek el = δik δjl 4.26 Show that δij δjk δki = n. 4.27 Write the determinantal array form of ǫij ǫkl outlining the pattern of the tensor indices in their relation to the indices of the rows and columns of the determinant array. 4.28 Prove the following identity using a truth table: ǫij ǫkl = δki δlj − δli δkj . 4.29 Prove the following identities justifying each step in your proofs: ǫil ǫkl = δki j i j ǫijk ǫlmk = δli δm − δm δl 4.30 Prove the following identities using other more general identities: ǫijk ǫljk = 2δli ǫijk ǫijk = 6 4.31 Outline the similarities and differences between the ordinary Kronecker delta tensor and the generalized Kronecker delta tensor. 4.32 Give the inductive mathematical definition of the generalized Kronecker delta tensor n δji11 ...i ...jn . n 4.33 Write the determinantal array form of the generalized Kronecker delta tensor δji11 ...i ...jn in terms of the ordinary Kronecker delta tensor. 4.34 Define ǫi1 ...in and ǫi1 ...in in terms of the generalized Kronecker delta tensor. ijk using an analytic or an inductive or a truth table 4.35 Prove the relation: ǫijk ǫlmn = δlmn method. 4.36 Demonstrate that the generalized Kronecker delta is an absolute tensor. mnq mn = δkl . 4.37 Prove the following relation justifying each step in your proof: δklq 4.38 Prove the common form of the epsilon-delta identity. 4.39 Prove the following generalization of the epsilon-delta identity: g ij ǫikl ǫjmn = gkm gln − gkn glm . 4.40 List and discuss all the main characteristics (e.g. symmetry) of the metric tensor. 4.41 How many types the metric tensor has? 4.42 Investigate the relation of the metric tensor of a given space to the coordinate systems of the space as well as its relation to the space itself by comparing the characteristics of the metric in different coordinate systems of the space such as being diagonal or not or having constant or variable components and so on. Hence, assess the status of the metric as a property of the space but with a form determined by the adopted coordinate system to describe the space and hence it is also a property of the coordinate system in this sense. 4.43 What is the relation between the covariant metric tensor and the length of an infinitesimal element of arc ds in a general coordinate system? 4.44 How the relation in question 4.43 will become (a) in an orthogonal coordinate system and (b) in an orthonormal Cartesian coordinate system? 4.7 Exercises and Revision 109 4.45 What is the characteristic feature of the metric tensor in orthogonal coordinate systems? 4.46 Write the mathematical expressions for the components of the covariant, contravariant and mixed forms of the metric tensor in terms of the covariant and contravariant basis vectors, Ei and Ei . 4.47 Write, in full tensor notation, the mathematical expressions for the components of the covariant and contravariant forms of the metric tensor, gij and g ij . 4.48 What is the relation between the mixed form of the metric tensor and the ordinary Kronecker delta tensor? 4.49 Explain why the metric tensor is not necessarily diagonal in general coordinate systems but it is necessarily symmetric. 4.50 Explain why the diagonal elements of the metric tensor in general coordinate systems are not necessarily of unit magnitude or positive but they are necessarily non-zero. 4.51 Explain why the mixed type metric tensor in any coordinate system is diagonal or in fact it is the unity tensor. 4.52 Show that the covariant and contravariant forms of the metric tensor, gij and g ij , are inverses of each other. 4.53 Why the determinant of the metric tensor should not vanish at any point in the space? 4.54 If the determinant of the covariant metric tensor gij is g, what is the determinant of the contravariant metric tensor g ij ? 4.55 Show that the metric tensor can be regarded as a transformation of the ordinary Kronecker delta tensor in its different variance types from an orthonormal Cartesian coordinate system to a general coordinate system. 4.56 Justify the use of the metric tensor as an index shifting operator using a mathematical argument. 4.57 Carry out the following index shifting operations recording the order of the indices: g ij Cklj gmn B nst gnl Dkmn 4.58 What is the difference between the three operations in question 4.57? 4.59 Why the order of the raised and lowered indices is important and hence it should be recorded? Mention one form of notation used to record the order of the indices. 4.60 What is the condition that should be satisfied by the metric tensor of a flat space? Give common examples of flat and curved spaces. 4.61 Considering a coordinate transformation, what is the relation between the determinants of the covariant metric tensor in the original and transformed coordinate systems, g and ḡ? 4.62 B is a “conjugate” or “associate” tensor of tensor A. What this means? 4.63 Complete and justify the following statement: “The components of the metric tensor are constants iff ...etc.”. 4.64 What are the covariant and absolute derivatives of the metric tensor? 4.65 Assuming an orthogonal coordinate system of an nD space, complete the following equations where the indexed g represents the metric tensor or its components, i 6= j 4.7 Exercises and Revision 110 in the second equation and there is no sum in the third equation:  gij =? g ii =? det g ij =? 4.66 Write, in matrix form, the covariant and contravariant metric tensor for orthonormal Cartesian, cylindrical and spherical coordinate systems. What distinguishes all these matrices? Explain and justify. 4.67 Referring to question 4.66, what is the relation between the diagonal elements of these matrices and the scale factors hi of the coordinates of these systems? 4.68 Considering the Minkowski metric, is the space of the mechanics of Lorentz transformations flat or curved? Is it homogeneous or not? What effect this can have on the length of element of arc ds? 4.69 Derive the following identities: gim ∂k g mj = −g mj ∂k gim ∂k gij = −gmj gni ∂k g nm 4.70 What is the dot product of A and B where A is a rank-2 covariant tensor and B is a contravariant vector? Write this operation in steps providing full justification of each step. 4.71 Derive an expression for the magnitude of a vector A when A is covariant and when A is contravariant. 4.72 Derive an expression for the cosine of the angle θ between two covariant vectors, A and B, and between two contravariant vectors C and D. 4.73 What is the meaning of the angle between two intersecting smooth curves? 4.74 What is the cross product of A and B where these are covariant vectors? 4.75 Complete the following equations assuming a general coordinate system of a 3D space: Ei × Ej =? Ei × Ej =? 4.76 Define the operations of scalar triple product and vector triple product of vectors using tensor language and assuming a general coordinate system of a 3D space. 4.77 What is the relation between the relative and absolute permutation tensors in their covariant and contravariant forms? 4.78 Define the determinant of a matrix B in tensor notation assuming a general coordinate system of a 3D space. 4.79 Derive the relation for the length of line element in general coordinate systems: (ds)2 = gij dui duj . How will this relation become when the coordinate system is orthogonal? Justify your answer. 4.80 Write the integral representing the length L of a t-parameterized space curve in terms of the metric tensor. 4.81 Using Eq. 295 plus the scale factors of Table 1, develop expressions for ds in orthonormal Cartesian, cylindrical and spherical coordinate systems. 4.82 Derive the following formula for the area of a differential element on the coordinate surface ui = constant in a 3D space assuming a general coordinate system: p dσ(ui = constant) = gg ii duj duk (i 6= j 6= k, no sum on i) 4.7 Exercises and Revision 111 How will this relation become when the coordinate system is orthogonal? 4.83 Using Eq. 301 plus the scale factors of Table 1, develop expressions for dσ on the coordinate surfaces in orthonormal Cartesian, cylindrical and spherical coordinate systems. 4.84 Derive the following formula for the volume of a differential element of a solid body in a 3D space assuming a general coordinate system: dτ = √ g du1 du2 du3 How will this relation become when the coordinate system is orthogonal? 4.85 Make a plot representing the volume of an infinitesimal element of a solid body in a 3D space as the magnitude of a scalar triple product of three vectors. 4.86 Use the expression of the volume element in general coordinate systems of nD spaces to find the formula for the volume element in orthogonal coordinate systems. 4.87 Using Eq. 304 plus the scale factors of Table 1, develop expressions for dτ in orthonormal Cartesian, cylindrical and spherical coordinate systems. Chapter 5 Tensor Differentiation Ordinary differentiation rules of partial and total derivatives do not satisfy the principle of invariance when they are applied to tensors in general coordinate systems due to the fact that the basis vectors in these systems are coordinate dependent. This means that the ordinary differentiation of non-scalar tensor components in general coordinates does not necessarily results in a tensor. Therefore, special types of differentiation should be defined so that when they apply to tensors they produce tensors. The essence of these operations is to extend the differentiation process to the basis vectors to which the tensor is referred and not only on the tensor components. The focus of this chapter is the operations of covariant and absolute differentiation which are closely linked. These operations represent generalization of ordinary differentiation in general coordinate systems with an objective of making the derivative of tensors comply with the principle of tensor invariance. Briefly, the differential change of a tensor in general coordinate systems is the result of a change in the basis vectors and a change in the tensor components. Hence, covariant and absolute differentiation, in place of the normal differentiation (i.e. partial and total differentiation respectively), are defined and employed to account for both of these changes by differentiating the basis vectors as well as the components of the tensors. Since the Christoffel symbols are crucial in the formulation and application of covariant and absolute differentiation, the first section of the present chapter is dedicated to these symbols and their properties. The subsequent two sections will then focus on the covariant differentiation and the absolute differentiation. 5.1 Christoffel Symbols We start by investigating the main properties of the Christoffel symbols which play crucial roles in tensor calculus in general and are needed for the subsequent development of the upcoming sections since they enter in the definition of covariant and absolute differentiation. The Christoffel symbols are classified as those of the first kind and those of the second kind. These two kinds are linked through the index raising and lowering operators although this does not mean they are general tensors (see next). Both kinds of Christoffel symbols are variable functions of coordinates since they depend in their definition on the metric tensor which is coordinate dependent in general. The Christoffel symbols of the first and second kind are not general tensors although they are affine tensors of rank-3. We note that affine tensors (see § 3.1.6) are tensors that correspond to admissible linear coordinate transformations from an original rectangular system of coordinates. As a consequence of not being tensors, if all the Christoffel symbols of either kind vanish in a particular coordinate system they will not necessarily vanish in 112 5.1 Christoffel Symbols 113 other coordinate systems (see § 3.1.4). For instance, all the Christoffel symbols of both kinds vanish in Cartesian coordinate systems but not in cylindrical or spherical coordinate systems, as will be established later in this section. The Christoffel symbols of the first kind are defined as: [ij, l] = 1 (∂j gil + ∂i gjl − ∂l gij ) 2 (307) where the indexed g is the covariant form of the metric tensor. The Christoffel symbols of the second kind are obtained by raising the third index of the Christoffel symbols of the first kind, and hence they are given by: Γkij = g kl [ij, l] = g kl (∂j gil + ∂i gjl − ∂l gij ) 2 (308) where the indexed g is the metric tensor in its contravariant and covariant forms with implied summation over l. Similarly, the Christoffel symbols of the first kind can be obtained from the Christoffel symbols of the second kind by reversing the above process through lowering the upper index, that is: l gkm Γkij = gkm g kl [ij, l] = δm [ij, l] = [ij, m] (309) where Eqs. 308, 220 and 172 are used. The Christoffel symbols of the first and second kind are symmetric in their paired indices, that is: [ij, k] = [ji, k] Γkij = Γkji (310) (311) These properties can be verified by shifting the indices in the definitions of the Christoffel symbols, as given by Eqs. 307 and 308, noting that the metric tensor is symmetric in its two indices (see § 4.5). For an nD space with n covariant basis vectors (E1 , E2 , . . . , En ) spanning the space, the partial derivative ∂j Ei for any given i and j is a vector within the space and hence it is in general a linear combination of all the basis vectors. The Christoffel symbols of the second kind are the components of this linear combination, that is: ∂j Ei = Γkij Ek (312) Similarly, for the contravariant basis vectors (E1 , E2 , . . . , En ) we have: ∂j Ei = −Γikj Ek (313) By inner product multiplication of Eq. 312 with Ek and Eq. 313 with Ek we obtain: Ek · ∂j Ei = +Γkij Ek · Ek = +Γkij δkk = +Γkij (314) 5.1 Christoffel Symbols 114 Ek · ∂j Ei = −Γikj Ek · Ek = −Γikj δkk = −Γikj (315) where Eqs. 49 and 172 are used. These equations reveal that the Christoffel symbols of the second kind are the projections of the partial derivative of the basis vectors in the direction of the basis vectors of the opposite variance type. From Eq. 314 and by using the index lowering operator, we obtain a similar relation that links the Christoffel symbols of the first kind to the basis vectors and their partial derivatives, that is: Ek · ∂j Ei = gmk Em · ∂j Ei = gmk Γm ij = [ij, k] (316) where Eq. 309 is used in the last step. This equation reveals that the Christoffel symbols of the first kind are the projections of the partial derivative of the covariant basis vectors in the direction of the basis vectors of the same variance type. The partial derivative of the components of the covariant metric tensor and the Christoffel symbols of the first kind satisfy the following identity: (317) ∂k gij = [ik, j] + [jk, i] This relation can be obtained from the partial derivative of the dot product of the basis vectors with the use of Eq. 316, that is: ∂k gij = ∂k (Ei · Ej ) = (∂k Ei ) · Ej + Ei · (∂k Ej ) = [ik, j] + [jk, i] (Eq. 213) (product rule) (Eq. 316) (318) We note that Eq. 317 is closely linked to the upcoming Ricci theorem (see § 5.2), that is: gij;k = ∂k gij −gaj Γaik −gia Γajk = ∂k gij −[ik, j]−[jk, i] = 0 ⇐⇒ ∂k gij = [ik, j]+[jk, i] (319) The relation given by Eq. 317 can also be written in terms of the Christoffel symbols of the second kind using the index shifting operator, that is: ∂k gij = gaj Γaik + gai Γajk (320) where Eq. 309 is used. Following the method of derivation given in Eq. 318, we obtain the following relation for the partial derivative of the components of the contravariant metric tensor:  (Eq. 214) (321) ∂k g ij = ∂k Ei · Ej   (product rule) = ∂ k Ei · Ej + Ei · ∂ k Ej   j j a i i a (Eq. 313) = −Γak E · E + E · −Γak E = −Γiak Ea · Ej − Γjak Ei · Ea = −g aj Γiak − g ia Γjak (Eq. 214) 5.1 Christoffel Symbols that is: 115 ∂k g ij = −g aj Γiak − g ia Γjak (322) g ij;k = ∂k g ij + g aj Γiak + g ia Γjak = 0 ⇐⇒ ∂k g ij = −g aj Γiak − g ia Γjak (323) Like Eq. 317 whose close link to the Ricci theorem is shown in Eq. 319, Eq. 322 can also be seen from this perspective, that is: The Christoffel symbols of the second kind with two identical indices of opposite variance type satisfy the following relations: Γjji = Γjij = 1 √ 1 1 √ ∂i g = ∂i (ln g) = ∂i (ln g) = √ ∂i g 2g 2 g (324) where the main relation can be derived as follows: Γjij = = = = = g jl (∂j gil + ∂i gjl − ∂l gij ) 2 g jl (∂l gij + ∂i gjl − ∂l gij ) 2 1 jl g ∂i gjl 2 1 jl gg ∂i gjl 2g 1 ∂i g 2g (Eq. 308 with k = j) (325) (relabeling dummy j, l in 1st term & g jl = g lj ) (derivative of determinant) All the other forms of Eq. 324 can be obtained from the derived form by simple algebraic manipulations with the use of the rules of differentiation and natural logarithms. In orthogonal coordinate systems, the Christoffel symbols of the first kind can be classified into three main groups, considering the identicality and difference of their indices, which are given by: 1 [ij, i] = [ji, i] = ∂j gii 2 1 [ii, j] = − ∂j gii 2 [ij, k] = 0 (no sum on i) (326) (i 6= j, no sum on i) (327) (i 6= j 6= k) (328) These three equations can be obtained directly from Eq. 307 with proper labeling of the indices (i.e. according to the labeling of the left hand side of Eqs. 326-328) noting that for orthogonal coordinate systems gij = 0 when i 6= j. For example, Eq. 327 can be obtained as follows: 1 (∂i gij + ∂i gij − ∂j gii ) 2 1 = (0 + 0 − ∂j gii ) 2 [ii, j] = (Eq. 307 with j → i and l → j) (gij = 0) (329) 5.1 Christoffel Symbols 116 1 = − ∂j gii 2 In fact, Eq. 326 can be obtained from Eq. 317 by exchanging j and k then replacing k with i followed by a simple algebraic manipulation. The middle equality of Eq. 326 is based on the fact that the Christoffel symbols are symmetric in their paired indices (Eq. 310). We note that Eq. 326 includes the case of j = i, i.e. when all the three indices are identical, and hence [ii, i] = 21 ∂i gii . For this reason, we did not add the condition i 6= j to Eq. 326. Returning to the above indicated consideration of the identicality and difference of the indices of the Christoffel symbols of the first kind, we have 4 main cases: (1) all the indices are identical, (2) only two non-paired indices are identical, (3) only the two paired indices are identical, and (4) all the indices are different. Eq. 326 represents case 1 and case 2, Eq. 327 represents case 3, and Eq. 328 represents case 4. This classification similarly applies to the Christoffel symbols of the second kind as represented by the upcoming Eqs. 331-333. In orthogonal coordinate systems where gij = g ij = 0 (i 6= j), the Christoffel symbols of the second kind are given by: Γijk = g ii [jk, i] [jk, i] = gii (Eq. 308) (330) (Eq. 235) with no sum on i. Hence, from Eqs. 326-328 and Eq. 330 the Christoffel symbols of the second kind in orthogonal coordinate systems are given by: g ii 1 1 = = ∂j gii = ∂j gii = ∂j ln gii 2 2gii 2 jj 1 g ∂j gii Γjii = − ∂j gii = − 2 2gjj Γijk = 0 Γiij Γiji (no sum on i) (331) (i 6= j, no sum on i or j) (332) (i 6= j 6= k) (333) where in the last step of Eq. 331 the well known rule of differentiating the natural logarithm is used. Like the Christoffel symbols of the first kind (refer to Eq. 326), Eq. 331 also includes the case of j = i, i.e. when all the three indices are identical, that is Γiii = 2g1ii ∂i gii . In orthogonal coordinate systems of a 3D space, the Christoffel symbols of both kinds vanish when the indices are all different, as seen above (Eqs. 328 and 333). Hence, out of a total of 27 symbols, representing all the possible permutations of the three indices including the repetitive ones, only 21 non-identically vanishing symbols are left since the six non-repetitive permutations are dropped. If we now consider that the Christoffel symbols are symmetric in their paired indices (Eqs. 310 and 311), then we are left with only 15 independent non-identically vanishing symbols since six other permutations representing these symmetric exchanges are also dropped because they are not independent. 5.1 Christoffel Symbols 117 Accordingly, in orthogonal coordinate systems in a 3D space the 15 independent nonidentically vanishing Christoffel symbols of the first kind are: [11, 1] = +h1 h1,1 [12, 1] = +h1 h1,2 [13, 3] = +h3 h3,1 [22, 3] = −h2 h2,3 [33, 1] = −h3 h3,1 [11, 2] = −h1 h1,2 [12, 2] = +h2 h2,1 [22, 1] = −h2 h2,1 [23, 2] = +h2 h2,3 [33, 2] = −h3 h3,2 [11, 3] = −h1 h1,3 [13, 1] = +h1 h1,3 [22, 2] = +h2 h2,2 [23, 3] = +h3 h3,2 [33, 3] = +h3 h3,3 (334) where the indices 1, 2, 3 stand for the three coordinates, h1 , h2 , h3 are the scale factors corresponding to these coordinates as defined previously (see § 2.5 and 2.6), and the comma indicates, as always, partial derivative with respect to the coordinate represented by the following index. For example, in cylindrical coordinates given by (ρ, φ, z), h2,1 means the partial derivative of h2 with respect to the first coordinate and hence h2,1 = ∂ρ ρ = 1 since h2 = ρ (refer to Table 1) and the first coordinate is ρ. Similarly, in spherical coordinates given by (r, θ, φ), h3,2 means the partial derivative of h3 with respect to the second coordinate and hence h3,2 = ∂θ (r sin θ) = r cos θ since h3 = r sin θ (refer to Table 1) and the second coordinate is θ. As indicated above, because the Christoffel symbols of the first kind are symmetric in their first two indices, the expression of the [21, 1] symbol for instance can be obtained from the expression of the [12, 1] symbol. The expressions given in Eq. 334 for the Christoffel symbols of the first kind in orthogonal coordinate systems are no more than simple applications of Eqs. 326 and 327 plus Eq. 235. For example, the entry [12, 1] can be obtained as follows: 1 [12, 1] = ∂2 g11 2 1 = ∂2 (h1 )2 2 = h1 ∂ 2 h1 = h1 h1,2 (Eq. 326) (335) (Eq. 235) (rules of differentiation) (notation) Similarly, in orthogonal coordinate systems in a 3D space the 15 independent nonidentically vanishing Christoffel symbols of the second kind are: h1,1 h1 h1,2 Γ112 = + h1 h 3,1 Γ313 = + h3 h2 h2,3 Γ322 = − (h3 )2 h3 h3,1 Γ133 = − (h1 )2 Γ111 = + h1 h1,3 (h3 )2 h1,3 =+ h1 h2,2 =+ h2 h3,2 =+ h3 h3,3 =+ h3 Γ211 = − h(h1 h1,2 )2 Γ311 = − Γ212 = + hh2,1 2 Γ113 Γ122 = − h(h2 h2,1 )2 1 Γ222 Γ223 = + hh2,3 2 Γ323 Γ233 = − h(h3 h3,2 )2 Γ333 2 2 (336) 5.1 Christoffel Symbols 118 where the symbols are as explained above. Again, since the Christoffel symbols of the second kind are symmetric in their lower indices, the non-vanishing entries which are not listed above can be obtained from the given entries by permuting the lower indices. The relations of Eq. 336 can be obtained from the relations of Eq. 334 by dividing by (hi )2 where i is the third index of the Christoffel symbol of the first kind. This can be justified by Eqs. 330 and 235. The relations of Eq. 336 can also be obtained directly from Eqs. 331 and 332 plus Eq. 235, as done for the Christoffel symbols of the first kind where Eqs. 326 and 327 were used. In any coordinate system, all the Christoffel symbols of the first and second kind vanish identically iff all the components of the metric tensor in the given coordinate system are constants. This can be seen from the definitions of the Christoffel symbols, as given by Eqs. 307 and 308, since the partial derivatives will vanish in this case. In affine coordinate systems all the components of the metric tensor are constants and hence all the Christoffel symbols of both kinds vanish identically. The prominent example is the orthonormal Cartesian coordinate systems where all the Christoffel symbols of the first and second kind are identically zero. This can also be seen from Eqs. 334 and 336 since the scale factors are constants (refer to Table 1). In cylindrical coordinate systems, identified by the coordinates (ρ, φ, z), all the Christoffel symbols of the first kind are zero except: [22, 1] = −ρ [12, 2] = [21, 2] = ρ (337) (338) where the indices 1, 2 stand for the coordinates ρ, φ respectively. Also, in cylindrical coordinate systems the non-zero Christoffel symbols of the second kind are given by: Γ122 = −ρ (339) 1 (340) Γ212 = Γ221 = ρ where the symbols are as explained above. Again, these results can be obtained from Eqs. 334 and 336 using the scale factors of Table 1. In spherical coordinate systems, identified by the coordinates (r, θ, φ), the non-zero Christoffel symbols of the first kind are given by: [22, 1] [33, 1] [12, 2] [33, 2] [13, 3] = −r = −r sin2 θ = [21, 2] = r = −r2 sin θ cos θ = [31, 3] = r sin2 θ [23, 3] = [32, 3] = r2 sin θ cos θ (341) (342) (343) (344) (345) (346) where the indices 1, 2, 3 stand for the coordinates r, θ, φ respectively. Also, in spherical coordinate systems the non-zero Christoffel symbols of the second kind are given by: Γ122 = −r (347) 5.1 Christoffel Symbols 119 Γ133 = −r sin2 θ 1 r = − sin θ cos θ 1 = Γ331 = r 3 Γ = cot θ = 32 Γ212 = Γ233 Γ313 Γ323 Γ221 = (348) (349) (350) (351) (352) where the symbols are as explained above. As before, these results can be obtained from Eqs. 334 and 336 using the scale factors of Table 1. As seen earlier (refer to § 2.2.2), all these coordinate systems (i.e. orthonormal Cartesian, cylindrical and spherical) are orthogonal systems, and hence Eqs. 334 and 336 do apply. Because there is an element of arbitrariness in the choice of the order of coordinates, and hence the order of their indices, the Christoffel symbols may be given in terms of coordinate symbols rather than their indices to be more explicit and to avoid ambiguity and confusion. For instance, in the above examples of the cylindrical coordinate systems identified by the coordinates (ρ, φ, z) we may use [φφ, ρ] instead of [22, 1] and use Γφρφ instead of Γ212 . Similarly, for the spherical coordinate systems identified by the coordinates (r, θ, φ) we may use [θθ, r] instead of [22, 1] and use Γφrφ instead of Γ313 . The Christoffel symbols of both kinds may also be subscripted or superscripted by the gij Γijk ) to reveal the symbol of the metric tensor of the given space (e.g. [ij, k]g and ij metric which the symbols are based upon. This is especially important when we have two or more different metrics related to two or more different spaces as it is the case, for instance, in differential geometry of 2D surfaces embedded in a 3D space where we have one metric for the surface and another metric for the 3D space. Alternatively, other means of distinction may be used such as using Latin or upper case indices for the Christoffel symbols of one metric and Greek or lower case indices for the Christoffel symbols of the other metric, e.g. [ij, k] and Γijk for the Christoffel symbols of the space metric and [αβ, γ] and Γγαβ for the Christoffel symbols of the surface metric. However, the latter methods do not apply when the indices are numeric rather than symbolic. The number of independent Christoffel symbols of each kind (first and second) in general coordinate systems is given by: n2 (n + 1) (353) NCI = 2 where n is the space dimension. The reason is that, due to the symmetry of the metric independent metric components, gij , and for each independent tensor there are n(n+1) 2 component there are n distinct Christoffel symbols. Alternatively, we have n3 permutations of the three indices including the repetitive ones, and out of these n3 permutations we have (n(n − 1)n) permutations whose paired indices are different where the three factors correspond to the first, second and third index respectively. Now, due to the symmetry of the Christoffel symbols in their paired indices, half of these (n(n − 1)n) permutations are identical to the other half and hence they are not independent. Therefore, the total 5.2 Covariant Differentiation 120 number of independent Christoffel symbols is: NCI = n3 − 2n3 − n3 + n2 n 3 + n2 n2 (n + 1) (n (n − 1) n) = = = 2 2 2 2 (354) The partial derivative of the Christoffel symbol of the first kind is given by: ∂k [ij, l] = 1 (∂k ∂j gil + ∂k ∂i gjl − ∂k ∂l gij ) 2 (355) where this equation is based on the definition of the Christoffel symbol of the first kind as given by Eq. 307. Finally, we should remark that there are several notations for the Christoffel symbols. As well as the symbols that we use in this book which may be the most common in use (i.e. [ij, k] for the first kind and Γkij for the second kind), the first kind may also     while the second kind may be symbolized as kij or ij be symbolized as Γijk or ij k k as well as other notations. There may be some advantages or disadvantages in these different notations. For example, the notations Γijk and Γkij may suggest, wrongly, that these symbols are tensors which is not the case since the Christoffel symbols are not general tensors although they are affine tensors. There may also be some advantages in the typesetting and writing of these notations or there are factors related to recognition, readability and even aesthetics. 5.2 Covariant Differentiation The focus of this section is the operation of covariant differentiation of tensors which is a generalization of the ordinary partial differentiation. The ordinary derivative of a tensor is not a tensor in general. The objective of covariant differentiation is to ensure the invariance of derivative (i.e. being a tensor) in general coordinate systems, and this results in applying more sophisticated rules using Christoffel symbols where different differentiation rules for covariant and contravariant indices apply. The resulting covariant derivative is a tensor which is one rank higher than the differentiated tensor. In brief, the covariant derivative is a partial derivative of the tensor that includes differentiating the basis vectors as well as differentiating the components, as we will see. Hence, the covariant derivative of a general tensor can be given generically by: ,m i,··· ,m j n j n ∂k (Ai,··· j,··· ,n Ei · · · Em E · · · E ) = Aj,··· ,n;k Ei · · · Em E · · · E (356) ,m where the expression of Ai,··· j,··· ,n;k will be given in the following paragraphs (see e.g. Eq. 366). More explicitly, the basis vectors in general curvilinear coordinate systems undergo changes in magnitude and direction as they move around in their own space, and hence they are functions of position. These changes should be accounted for when calculating the derivatives of non-scalar tensors in such general systems. Therefore, terms based on using Christoffel symbols are added to the ordinary derivative terms to correct for these changes 5.2 Covariant Differentiation 121 and this more comprehensive form of derivative is called the covariant derivative. Since in rectilinear coordinate systems the basis vectors are constants, the Christoffel symbol terms vanish identically and hence the covariant derivative reduces to the ordinary partial derivative, but in the other coordinate systems these terms are present in general. As a consequence, the ordinary derivative of a non-scalar tensor is a tensor iff the coordinate transformations from Cartesian systems to that system are linear. It has been suggested that the “covariant” label is an indication that the covariant differentiation operator ∇;i is in the covariant position. However, it may also be true that “covariant” means “invariant” as the term “covariant” is also used in tensor calculus to mean “invariant”. In fact, “covariant” as opposite to “contravariant” applies even to the common form of ordinary partial differentiation since the commonly used partial differential operator ∂i is also in the covariant position. Anyway, contravariant differentiation, which may be associated with the operator ∇;j , can also be defined for covariant and contravariant tensors by raising the differentiation index of the covariant derivative using the index raising operator, e.g. Ai;j = g jk Ai;k Ai;j = g jk Ai ;k (357) However, contravariant differentiation is rarely used. As an example of how to obtain the covariant derivative of a tensor, let have a vector A represented by contravariant components in general curvilinear coordinates, that is: A = Ai Ei . We differentiate this vector following the normal rules of differentiation and taking account of the fact that the basis vectors in general curvilinear coordinate systems are differentiable functions of position and hence they, unlike their rectilinear counterparts, are subject to differentiation using the product rule, that is: (definition) A;j = ∂j A i = ∂ j A Ei  = Ei ∂j Ai + Ai ∂j Ei (product rule) = Ei ∂j Ai + Ai Γkij Ek (Eq. 312) = = i k Ei ∂j A + A Γikj Ei  ∂j Ai + Ak Γikj Ei = Ai;j Ei (358) (relabeling dummy indices i ↔ k) (taking common factor) (notation) where Ai;j , which is a rank-2 mixed tensor, is labeled the “covariant derivative” of Ai . Similarly, for a vector represented by covariant components in general curvilinear coordinate systems, A = Ai Ei , we have: (definition) A;j = ∂j A = ∂j Ai E i  = Ei ∂j Ai + Ai ∂j Ei (product rule) = Ei ∂j Ai − Ai Γikj Ek (Eq. 313) (359) 5.2 Covariant Differentiation 122 = Ei ∂j Ai − Ak Γkij Ei  = ∂j Ai − Ak Γkij Ei = Ai;j E (relabeling dummy indices i ↔ k) (taking common factor) i (notation) The same rules apply to tensors of higher ranks. For example, for a rank-2 mixed tensor, A = Ai j Ei Ej , we have: (definition) A;k = ∂k A j = ∂ k Ai E i E j  = ∂k Ai j Ei Ej  = ∂k Ai j Ei Ej  = ∂k Ai j Ei Ej  = ∂k Ai j Ei Ej   (product rule) + Ai j ∂k Ei Ej + Ai j Ei (∂k Ej )   j i j a a i + Ai −Γak E Ej + Ai E Γjk Ea (Eqs. 312 & 313) − Ai j Γiak Ea Ej + Ai j Γajk Ei Ea − Aaj Γaik Ei Ej + Ai a Γjak Ei Ej  = ∂k Ai j − Aaj Γaik + Ai a Γjak Ei Ej = Ai j ;k E i (360) Ej (Eq. 125) (relabeling dummy indices) (taking common factor) (notation) Based on the above arguments and examples, the main rules of covariant differentiation can be outlined in the following examples. For a differentiable vector A, the covariant derivative of the covariant and contravariant forms of the vector is given by: Aj;i = ∂i Aj − Γkji Ak Aj;i = ∂i Aj + Γjki Ak (covariant) (361) (contravariant) (362) Similarly, for a differentiable rank-2 tensor A, the covariant derivative of the covariant, contravariant and mixed forms of the tensor is given by: Ajk;i = ∂i Ajk − Γlji Alk − Γlki Ajl Ajk;i Akj;i jk = ∂i A + = ∂i Akj + Γjli Alk Γkli Alj + Γkli Ajl − Γlji Akl (covariant) (363) (contravariant) (364) (mixed) (365) We note that for the mixed form there are two possibilities: one associated with the dyad Ej Ek and the other with the dyad Ek Ej . Following the methods and techniques outlined in the previous examples, we can easily deduce the pattern of the operation of covariant differentiation. To obtain the covariant derivative of a tensor in general, we start with an ordinary partial derivative term of the component of the given tensor. Then for each tensor index an extra Christoffel symbol term is added, positive for contravariant indices and negative for covariant indices, where the differentiation index is one of the lower indices in the Christoffel symbol of the second kind. Hence, for a differentiable rank-n tensor A in general coordinate systems the covariant derivative with respect to the q th coordinate is given by: ij...k aj...k ij...a i j ia...k k Aij...k lm...p;q = ∂q Alm...p + Γaq Alm...p + Γaq Alm...p + · · · + Γaq Alm...p (366) 5.2 Covariant Differentiation 123 ij...k ij...k a a −Γalq Aij...k am...p − Γmq Ala...p − · · · − Γpq Alm...a In fact, there is practically only one possibility for the arrangement and labeling of the indices in the Christoffel symbol terms of the covariant derivative of a tensor of any rank if the following rules are observed: 1. The second subscript index of the Christoffel symbol is the differentiation index. 2. The differentiated index of the tensor in the Christoffel symbol term is contracted with one of the indices of the Christoffel symbol using a new label and hence they are opposite in their covariant and contravariant type. 3. The label of the differentiated index is transferred from the tensor to the Christoffel symbol keeping its position as covariant or contravariant. 4. All the other indices of the tensor in the concerned Christoffel symbol term keep their labels, position and order. The ordinary partial derivative term in the covariant derivative expressions (see e.g. Eq. 366) represents the rate of change of the tensor components with change of position as a result of moving along the coordinate curve of the differentiation index, while the Christoffel symbol terms represent the change experienced by the local basis vectors as a result of the same movement. This can be seen from the development of Eqs. 358-360 where the indicated terms correspond to the components and basis vectors according to the product rule of differentiation. From the above discussion, it is obvious that to obtain the covariant derivative, the Christoffel symbols of the second kind should be obtained and these symbols are dependent on the metric tensor. Hence, the covariant derivative is dependent on having the space metric corresponding to the particular coordinate system. We also note that the covariant derivative of a tensor is a tensor whose covariant rank is higher than the covariant rank of the original tensor by one. Hence, the covariant derivative of a rank-n tensor of type (r, s) is a rank-(n + 1) tensor of type (r, s + 1). In all coordinate systems, the covariant derivative of a differentiable scalar function of position, f , is the same as the ordinary partial derivative, that is: f;i = f,i = ∂i f (367) This is justified by the fact that the covariant derivative is different from the ordinary partial derivative because the basis vectors in general curvilinear coordinate systems are dependent on their spatial position, and since a scalar is independent of the basis vectors the covariant derivative and the partial derivative are identical. The derivation of the expressions of the covariant derivative of contravariant and covariant vectors and higher rank tensors, as given by Eqs. 358-360, clearly justifies this logic where the product rule does not apply due to the absence of a basis vector in the representation of a scalar. This can also be concluded from the covariant derivative rules as demonstrated in the previous statements and formulated in the above equations like Eq. 366 since a scalar has no free index and hence it cannot have any Christoffel symbol term. By a similar reasoning, since the Christoffel symbols are identically zero in rectilinear coordinate systems, the covariant derivative in these systems is the same as the ordinary partial derivative for all tensor 5.2 Covariant Differentiation 124 ranks, whether scalars or not. This can also be seen from the product rule (as employed in the development of Eqs. 358-360 for instance) where all the terms involving differentiation of basis vectors will vanish since these vectors are constant in rectilinear systems. Another important fact about covariant differentiation is that the covariant derivative of the metric tensor in its covariant, contravariant and mixed forms is zero in all coordinate systems and hence it is treated like a constant with respect to covariant differentiation. Accordingly, the covariant derivative operator bypasses the metric tensor, e.g.  (368) ∂;m gij Aj = gij ∂;m Aj and hence the metric tensor commutes with the covariant differential operator. We will expand on this issue later in this section. Several rules of ordinary differentiation are naturally extended to covariant differentiation. For example, covariant differentiation is a linear operation with respect to algebraic sums of tensor terms and hence the covariant derivative of a sum is the sum of the covariant derivatives of the terms, that is:[22] (aA ± bB);i = aA;i ± bB;i (369) where a and b are scalar constants and A and B are differentiable tensors. The product rule of ordinary differentiation also applies to the covariant differentiation of inner and outer products of tensors, that is: (A ◦ B);i = A;i ◦ B + A ◦ B;i (370) where the symbol ◦ denotes an inner or outer product operator. However, as seen in this equation, the order of the tensors should be observed since tensor multiplication, unlike ordinary algebraic multiplication, is not commutative (refer to § 3.2.3). The product rule is valid for the inner product of tensors because the inner product is an outer product operation followed by a contraction of indices, and covariant differentiation and contraction of indices do commute as we will see later. A principal difference between partial differentiation and covariant differentiation is that for successive differential operations with respect to different indices the ordinary partial derivative operators do commute with each other, assuming that the well known continuity condition is satisfied, but the covariant differential operators do not commute, that is: ∂i ∂j = ∂j ∂i ∂;i ∂;j 6= ∂;j ∂;i (371) This will be verified later in this section. As indicated earlier, according to the “Ricci theorem” the covariant derivative of the covariant, contravariant and mixed forms of the metric tensor is zero. This has nothing to do with the metric tensor being a constant function of coordinates, which is true only for [22] We use a semicolon with symbolic notation of tensors in this equation and other similar equations for the sake of clarity; the meaning should be obvious. 5.2 Covariant Differentiation 125 rectilinear coordinate systems, but this arises from the fact that the covariant derivative quantifies the change with position of the basis vectors in magnitude and direction as well as the change in components, and these contributions in the case of the metric tensor cancel each other resulting in a total null effect. For example, for the covariant form of the metric tensor we have:  g;k = ∂k gij Ei Ej (definition) (372)  j  i i i j j = (∂k gij ) E E + gij ∂k E E + gij E ∂k E (product rule)   j j l i i l i j (Eqs. 317 & 313) = ([ik, j] + [jk, i]) E E + gij −Γlk E E + gij E −Γlk E = [ik, j] Ei Ej + [jk, i] Ei Ej − gij Γilk El Ej − gij Γjlk Ei El = [ik, j] Ei Ej + [jk, i] Ei Ej − [lk, j] El Ej − [lk, i] Ei El i j i j i j i = [ik, j] E E + [jk, i] E E − [ik, j] E E − [jk, i] E E =0 j (Eq. 309) (relabeling dummy l) Similarly, for the contravariant form of the metric tensor we have:  (definition) (373) g;k = ∂k g ij Ei Ej  (product rule) = ∂k g ij Ei Ej + g ij (∂k Ei ) Ej + g ij Ei (∂k Ej )   j (Eqs. 322 & 312) = −g aj Γiak − g ia Γak Ei Ej + g ij (Γaik Ea ) Ej + g ij Ei Γajk Ea = −g aj Γiak Ei Ej − g ia Γjak Ei Ej + g ij Γaik Ea Ej + g ij Γajk Ei Ea = −g aj Γiak Ei Ej − g ia Γjak Ei Ej + g aj Γiak Ei Ej + g ia Γjak Ei Ej =0 As for the mixed form of the metric tensor we have:  g;k = ∂k δ ij Ei Ej   = ∂k δ ij Ei Ej + δ ij (∂k Ei ) Ej + δ ij Ei ∂k Ej  = 0 + δ ij (Γaik Ea ) Ej + δ ij Ei −Γjak Ea = = δ ij Γaik Ea Ej − δ ij Γjak Ei Ea Γajk Ea Ej − Γiak Ei Ea (relabeling indices) (definition) (374) (product rule) (Eqs. 312 & 313) (Eq. 172) = Γijk Ei Ej − Γijk Ei Ej (relabeling dummy indices) =0 As a result of the Ricci theorem, the metric tensor behaves as a constant with respect to the covariant derivative operation, that is: g;k = 0 (375) where g is the metric tensor in its covariant or contravariant or mixed form, as seen above. Consequently, the covariant derivative operator bypasses the metric tensor, that is: (g ◦ A);k = g ◦ A;k (376) 5.2 Covariant Differentiation 126 where A is a general tensor and the symbol ◦ denotes an inner or outer tensor product.[23] The commutativity of the covariant derivative operator and the metric tensor acting as an index shifting operator may be demonstrated more vividly by using the indicial notation, e.g.   (377) gik Ak;j = gik Ak;j = Ai;j = (Ai );j = gik Ak ;j  (378) gmi gnj Amn;k = gmi gnj Amn;k = Aij;k = (Aij );k = (gmi gnj Amn );k where we see that the sequence of these operations, seen in one order from the right and in another order from the left, has no effect on the final result in the middle. Like the metric tensor, the ordinary Kronecker delta tensor is constant with regard to covariant differentiation and hence the covariant derivative of the Kronecker delta is identically zero, that is:[24] i δj;k = ∂k δji + δja Γiak − δai Γajk = 0 + δja Γiak − δai Γajk = 0 + Γijk − Γijk (Eq. 365) (379) (δji is constant) (Eq. 172) =0 Accordingly, the covariant differential operator bypasses the Kronecker delta tensor which is involved in inner and outer tensor products:[25] (δ ◦ A);k = δ ◦ A;k (380) The rule of the Kronecker delta may be regarded as an instance of the rule of the metric tensor, as stated by the Ricci theorem, since the Kronecker delta is a metric tensor for certain systems and types. Like the ordinary Kronecker delta, the covariant derivative of the generalized Kronecker delta is also identically zero. This may be deduced from Eq. 379 plus Eq. 199 where the generalized Kronecker delta is given as a determinant consisting of ordinary Kronecker deltas and hence it is a sum of products of ordinary Kronecker deltas whose partial derivative vanishes because each term in the derivative contains a derivative of an ordinary Kronecker delta. Based on the previous statements, we conclude that covariant differentiation and contraction of index operations commute with each other, e.g.    k ij k ij k ik ik (381) Aij k;l δj = Ak;l δj = Ak;l = Ak ;l = Ak δj ;l where we see again that the different sequences from the right and from the left produce the same result in the middle. For clarity, we represented the contraction operation in this [23] Although the metric tensor is normally used in inner product operations for raising and lowering indices, the possibility of its involvement in outer product operations should not be ruled out. [24] For diversity, we use indicial notation rather than symbolic notation which we used for example in verifying g;k = 0. [25] Like the metric tensor, the Kronecker delta tensor is normally used in inner product operations for replacement of indices; however the possibility of its involvement in outer product operations should not be ruled out. 5.2 Covariant Differentiation 127 example by an inner product operation of the ordinary Kronecker delta tensor with the contracted tensor. For a differentiable function f (x, y) of class C 2 (i.e. all the second order partial derivatives of the function do exist and are continuous), the mixed partial derivatives are equal, that is: ∂x ∂y f = ∂y ∂x f (382) However, even if the components of a tensor satisfy this condition (i.e. being of class C 2 ), this is not sufficient for the equality of the mixed covariant derivatives. What is required for the mixed covariant derivatives to be equal is the vanishing of the Riemann-Christoffel curvature tensor (see § 7.2.1) which is equivalent to having an intrinsically flat space. This will be verified later in this section. Higher order covariant derivatives are defined as derivatives of derivatives by successive repetition of the process of covariant differentiation. However, the order of differentiation, in the case of differentiating with respect to different indices, should be respected as stated in the previous statements. For example, the second order mixed jk covariant derivative of a contravariant vector A is given by:  (383) Ai;jk = Ai;j ;k = ∂k Ai;j + Γiak Aa;j − Γajk Ai;a    = ∂k ∂j Ai + Γiaj Aa + Γiak ∂j Aa + Γabj Ab − Γajk ∂a Ai + Γiba Ab = ∂k ∂ j Ai + Γiaj ∂ k Aa + Aa ∂k Γiaj + Γiak ∂j Aa + Γiak Γabj Ab − Γajk ∂a Ai − Γajk Γiba Ab that is: Ai;jk = ∂k ∂ j Ai + Γiaj ∂ k Aa − Γajk ∂a Ai + Γiak ∂j Aa + Aa ∂k Γiaj − Γbjk Γiba + Γibk Γbaj  (384) The second order mixed kj covariant derivative of a contravariant vector can be derived similarly. However, it can be obtained more easily from the last equation by interchanging the j and k indices, that is:  (385) Ai;kj = ∂j ∂ k Ai + Γiak ∂ j Aa − Γakj ∂a Ai + Γiaj ∂k Aa + Aa ∂j Γiak − Γbkj Γiba + Γibj Γbak The inequality of the jk and kj mixed derivatives in general can be verified by subtracting the two sides of the last two equations from each other where the right hand side will not vanish. Similarly, the second order mixed jk covariant derivative of a covariant vector A is given by: Ai;jk = (Ai;j );k = ∂k Ai;j − Γaik Aa;j − Γajk Ai;a    = ∂k ∂j Ai − Γbij Ab − Γaik ∂j Aa − Γbaj Ab − Γajk ∂a Ai − Γbia Ab (386) = ∂k ∂j Ai − Γbij ∂k Ab − Ab ∂k Γbij − Γaik ∂j Aa + Γaik Γbaj Ab − Γajk ∂a Ai + Γajk Γbia Ab 5.2 Covariant Differentiation 128 that is: Ai;jk = ∂k ∂j Ai − Γaij ∂k Aa − Γaik ∂j Aa − Γajk ∂a Ai − Aa ∂k Γaij − Γbik Γabj − Γbjk Γaib  (387) The second order mixed kj covariant derivative of a covariant vector can be obtained from the last equation by interchanging the j and k indices, that is:  (388) Ai;kj = ∂j ∂k Ai − Γaik ∂j Aa − Γaij ∂k Aa − Γakj ∂a Ai − Aa ∂j Γaik − Γbij Γabk − Γbkj Γaib Again, the inequality of the jk and kj mixed derivatives can be verified by subtracting the two sides of the last two equations from each other where the right hand side will not vanish. We will see in § 7.2.1 that the mixed second order covariant derivatives of a covariant vector Ai are linked through the Riemann-Christoffel curvature tensor Raijk by the following relation: Ai;jk − Ai;kj = Aa Raijk (389) This relation can be verified by subtracting the two sides of Eq. 388 from the two sides of Eq. 387 and employing the definition of the Riemann-Christoffel curvature tensor of Eq. 560 to the right hand side, that is:   Ai;jk − Ai;kj = −Aa ∂k Γaij − Γbik Γabj − Γbjk Γaib + Aa ∂j Γaik − Γbij Γabk − Γbkj Γaib (390)   = Aa ∂j Γaik − Γbij Γabk − Γbkj Γaib − Aa ∂k Γaij − Γbik Γabj − Γbjk Γaib  = Aa ∂j Γaik − Γbij Γabk − Γbkj Γaib − ∂k Γaij + Γbik Γabj + Γbjk Γaib  = Aa ∂j Γaik − Γbij Γabk − ∂k Γaij + Γbik Γabj = Aa Raijk where the last step is based on Eq. 560. The covariant derivatives of relative tensors, which are also relative tensors of the same weight as the original tensors, are obtained by adding a weight term to the normal formulae of covariant derivative. Hence, the covariant derivative of a relative scalar with weight w is given by: f;i = f,i − wf Γjji (391) while the covariant derivative of relative tensors of higher ranks with weight w is obtained by adding the following term to the right hand side of Eq. 366: a − wAij...k lm...p Γaq (392) Unlike ordinary differentiation, the covariant derivative of a non-scalar tensor with constant components is not zero in general curvilinear coordinate systems due to the presence of the Christoffel symbols in the definition of the covariant derivative, as given by Eq. 366. More explicitly, even though the partial derivative term is zero because the components are constant, the Christoffel symbol terms are not zero in general because the basis vectors are variables (refer to the derivation in Eqs. 358-360). 5.3 Absolute Differentiation 129 In rectilinear coordinate systems, the Christoffel symbols are identically zero because the basis vectors are constants, and hence the covariant derivative is the same as the ordinary partial derivative for all tensor ranks. As a result, when the components of the metric tensor gij are constants, as it is the case for example in rectangular coordinate systems, the covariant derivative becomes ordinary partial derivative. The constancy of the metric tensor in this case can be concluded from the definition of the components of the metric tensor as dot products of the basis vectors, as seen for example in Eqs. 213 and 214 (also see Table 1). We remark that for a differentiable covariant vector A which is a gradient of a scalar field we have: Ai;j = Aj;i (393) This can be easily verified by defining A as: Ai = f,i where f is a scalar and hence we have: (394) Ai;j = (f,i );j = (f,i ),j − Γkij f,k (Eq. 361) = f,ij − Γkij f,k = f,ji − Γkji f,k (Eqs. 382 & 311) (Eq. 361) = (f,j );i = Aj;i Another important remark is that the covariant derivative of the basis vectors of the covariant and contravariant types is identically zero, that is: Ei;j = ∂j Ei − Γkij Ek = +Γkij Ek − Γkij Ek = 0 Ei;j = ∂j Ei + Γikj Ek = −Γikj Ek + Γikj Ek = 0 (395) (396) where Eqs. 361 and 362 are used in the first steps, while Eqs. 312 and 313 are used in the second steps. 5.3 Absolute Differentiation The absolute derivative of a tensor along a t-parameterized curve C(t) in an nD space with respect to the parameter t is the inner product of the covariant derivative of the tensor and the tangent vector to the curve. In brief, the absolute derivative is a covariant derivative of a tensor along a curve. For a tensor Ai , the inner product of Ai;j , which is i a tensor, with another tensor is a tensor. Now, if the other tensor is du , which is the dt tangent vector to a t-parameterized curve C(t) given by the equations ui = ui (t) where i = 1, · · · , n, then the inner product: Ai;r dur dt (397) 5.3 Absolute Differentiation 130 is a tensor of the same rank and type as the tensor Ai . The tensor given by the expression of Eq. 397 is called the “absolute” or “intrinsic” or “absolute covariant” derivative of the i , that is: tensor Ai along the curve C and is symbolized by δA δt δAi dur = Ai;r δt dt (398) In fact, instead of basing the definition of absolute differentiation on the definition of the covariant differentiation as a dot product of the covariant derivative with the tangent to the curve as seen above, we can define absolute differentiation independently by following the same method that we used to obtain the covariant derivative through applying the differentiation process on the basis vectors as well as the components of the tensor (as seen in the derivation of Eqs. 358-360) and hence the expressions of the absolute derivative (e.g. the upcoming Eqs. 401 and 402) can be obtained directly by applying total differentiation to the tensor including its basis vectors. For example, the absolute derivative of a contravariant vector A = Ai Ei can be obtained as follows:  d dA Ai Ei = dt dt dAi dEi + Ai = Ei dt dt i dA ∂Ei duj = Ei + Ai j dt ∂u dt dAi duj = Ei + Ai Γkij Ek dt dt i duj dA + Ei Ak Γikj = Ei dt  dti  j dA du = Ei + Ak Γikj dt dt δAi Ei = δt (399) (product rule) (chain rule) (Eq. 312) (exchanging dummy indices i, k) (definition of intrinsic derivative) which is the same as the upcoming Eq. 401. Following the above definitions, the absolute derivative of a differentiable scalar f is the same as the ordinary total derivative, that is: δf df = δt dt (400) This is because the covariant derivative of a scalar is the same as the ordinary partial derivative or because a scalar has no association with basis vectors to differentiate. The absolute derivative of a differentiable contravariant vector Ai with respect to the parameter t is given by: δAi dAi duj = + Γikj Ak (401) δt dt dt 5.3 Absolute Differentiation 131 Similarly, for a differentiable covariant vector Ai we have: dAi duj δAi = − Γkij Ak δt dt dt (402) Absolute differentiation can be easily extended to higher rank (> 1) differentiable tensors of type (m, n) along parameterized curves following the given examples and the pattern of covariant derivative. For instance, the absolute derivative of a mixed tensor of type (1, 2) Aijk along a t-parameterized curve C is given by: dAijk δAijk dub dub dub dub = Aijk;b = + Γiab Aajk − Γajb Aiak − Γabk Aija δt dt dt dt dt dt (403) Since the absolute derivative is given generically by:[26] δA duk = A;k δt dt (404) it can be seen as an instance of the chain rule of differentiation where the two contracted indices represent the in-between coordinate differential. This can also be concluded from the above method of derivation of Eq. 399. Because the absolute derivative along a curve is just an inner product of the covariant derivative with the tangent vector to the curve, the well known rules of ordinary differentiation of sums and products also apply to absolute differentiation, as for covariant differentiation, that is: δ δA δB (aA + bB) = a +b δt δt δt     δ δB δA (A ◦ B) = ◦B + A◦ δt δt δt (405) (406) where a and b are constant scalars, A and B are differentiable tensors and the symbol ◦ denotes an inner or outer product of tensors. However, the order of the tensors in the products should be observed since tensor multiplication is not commutative. Because absolute differentiation follows the style of covariant differentiation, the metric tensor in its different variance types is in lieu of a constant with respect to absolute differentiation, that is: δgij δg ij =0 =0 (407) δt δt and hence it passes through the absolute derivative operator, that is: δAj δ (gij Aj ) = gij δt δt [26] δ (g ij Aj ) δAj = g ij δt δt (408) We use the absolute derivative notation with the symbolic notation of tensors in this equation and some of the upcoming equations to ease the notation. This is similar to the use of the covariant derivative notation with the symbolic notation as seen here and in previous equations. The meaning of these notations should be clear 5.3 Absolute Differentiation 132 For coordinate systems in which all the components of the metric tensor are constants, the absolute derivative is the same as the ordinary total derivative, as it is the case in rectilinear coordinate systems, because the Christoffel symbol terms are zero in these systems according to Eq. 308. The absolute derivative of a tensor along a given curve is unique, and hence the ordinary derivative of the tensor along that curve in a rectangular coordinate system is the same as the absolute derivative of the tensor along that curve in any other system although the forms in the two systems may be different. To sum up, we list in the following bullet points the main rules of covariant and absolute differentiation: 1. Tensor differentiation (represented by covariant and absolute differentiation) is the same as ordinary differentiation (represented by partial and total differentiation) but with the application of the differentiation process not only on the tensor components but also on the basis vectors that associate these components. 2. The sum and product rules of differentiation apply to covariant and absolute differentiation as for ordinary differentiation (i.e. partial and total). 3. The covariant and absolute derivatives of tensors are tensors. 4. The covariant and absolute derivatives of scalars and affine tensors of higher ranks are the same as the ordinary derivatives. 5. The covariant and absolute derivative operators commute with the contraction of indices. 6. The covariant and absolute derivatives of the metric, Kronecker and permutation tensors as well as the basis vectors vanish identically in any coordinate system, that is: gij|q = 0 g ij|q = 0 (409) i δj|q = 0 ij = 0 δkl|q (410) ǫijk|q = 0 ǫijk |q = 0 (411) Ei|q = 0 Ei|q = 0 (412) where the sign | represents covariant or absolute differentiation with respect to the space coordinate uq . Hence, these tensors should be treated like constants in tensor differentiation. This applies to the covariant and contravariant forms of these tensors, as well as the mixed form when it is applicable. It also applies to the generalized, as well as the ordinary, Kronecker delta as seen in the above equations. 7. Covariant differentiation increases the covariant rank of the differentiated tensor by one, while absolute differentiation does not change the rank or type of the differentiated tensor. 8. The covariant and absolute derivatives in rectilinear systems are the same as the partial and total derivatives respectively for all tensor ranks. 5.4 Exercises and Revision 5.4 133 Exercises and Revision 5.1 Why tensor differentiation (represented by covariant and absolute derivatives) is needed in general coordinate systems to replace the ordinary differentiation (represented by partial and total derivatives)? 5.2 Show that in general coordinate systems, the ordinary differentiation of the components of non-scalar tensors with respect to the coordinates will not produce a tensor in general. 5.3 “The Christoffel symbols are affine tensors but not tensors”. Explain and justify this statement. 5.4 What is the difference between the first and second kinds of the Christoffel symbols? 5.5 Show that the Christoffel symbols of both kinds are not general tensors by giving examples of these symbols being vanishing in some systems but not in other systems and considering the universality of the zero tensor (see § 3.1.4). 5.6 State the mathematical definitions of the Christoffel symbols of the first and second kinds. How these two kinds are transformed from each other? 5.7 What is the significance of the Christoffel symbols being solely dependent on the coefficients of the metric tensor in their relation to the underlying space and coordinate system? 5.8 Do the Christoffel symbols represent a property of the space, a property of the coordinate system, or a property of both? 5.9 If some of the Christoffel symbols vanish in a particular curvilinear coordinate system, should these some necessarily vanish in other curvilinear coordinate systems? Justify your answer by giving some examples. 5.10 Verify that the Christoffel symbols of the first and second kind are symmetric in their paired indices by using their mathematical definitions. 5.11 Correct, if necessary, the following equations: ∂j Ei = −Γkij Ek ∂j Ei = −Γimj Em 5.12 What is the significance of the following equations? Ek · ∂j Ei = Γkij Ek · ∂j Ei = −Γikj Ek · ∂j Ei = [ij, k] 5.13 Derive the following relations giving full explanation of each step: √ ∂j gil = [ij, l] + [lj, i] Γjji = ∂i (ln g) 5.14 Assuming an orthogonal coordinate system, verify the following relation: [ij, k] = 0 where i 6= j 6= k. 5.15 Assuming an orthogonal coordinate system, verify the following relation: Γiji = 1 ∂ ln gii with no sum over i. 2 j 5.16 Considering the identicality and difference of the indices of the Christoffel symbols of either kind, how many cases we have? List these cases. 5.17 Prove the following relation which is used in Eq. 325: ∂i g = gg jl ∂i gjl . 5.4 Exercises and Revision 134 5.18 In orthogonal coordinate systems of a 3D space the number of independent nonidentically vanishing Christoffel symbols of either kind is only 15. Explain why. 5.19 Verify the following equations related to the Christoffel symbols in orthogonal coordinate systems in a 3D space: [12, 1] = h1 h1,2 Γ323 = h3,2 h3 5.20 Justify the following statement: “In any coordinate system, all the Christoffel symbols of either kind vanish identically iff all the components of the metric tensor in the given coordinate system are constants”. 5.21 Using Eq. 307 with Eq. 239, find the Christoffel symbols of the first kind corresponding to the Euclidean metric of cylindrical coordinate systems. 5.22 Give all the Christoffel symbols of the first and second kind of the following coordinate systems: orthonormal Cartesian, cylindrical and spherical. 5.23 Mention two important properties of the Christoffel symbols of either kind with regard to the order and similarity of their indices. 5.24 Using the entries in Eq. 336 and Table 1 and the properties of the Christoffel symbols of the second kind, derive these symbols corresponding to the metrics of the coordinate systems of question 5.22. 5.25 Write the following Christoffel symbols in terms of the coordinates instead of the indices assuming a cylindrical system: [12, 1], [23, 1], Γ221 and Γ332 . Do the same assuming a spherical system. 5.26 Show that all the Christoffel symbols will vanish when the components of the metric tensor are constants. 5.27 Why the Christoffel symbols of either kind may be superscripted or subscripted by the symbol of the underlying metric tensor? When this (or other measures for indicating the underlying metric tensor) becomes necessary? Mention some of the other measures used to indicate the underlying metric tensor. 5.28 Explain why the total number of independent Christoffel symbols of each kind is equal 2 to n (n+1) . 2 5.29 Why covariant differentiation of tensors is regarded as a generalization of the ordinary partial differentiation? 5.30 In general curvilinear coordinate systems, the variation of the basis vectors should also be considered in the differentiation process of non-scalar tensors. Why? 5.31 State the mathematical definition of contravariant differentiation of a tensor Ai . 5.32 Obtain analytical expressions for Ai;j and B i;j by differentiating the vectors A = Ai Ei and B = B i Ei . 5.33 Repeat question 5.32 with the rank-2 tensors C = Cij Ei Ej and D = Dij Ei Ej to obtain Cij;k and Dij;k . 5.34 For a differentiable tensor A of type (m, n), the covariant derivative with respect to the coordinate uk is given by: i i ...i m Aji11ij22...i ...jn ;k = ∂Aj1 j2 ...jm n 1 2 ∂uk im i1 i2 ...l m m + Γilk2 Aji11l...i +Γilk1 Ajli12j...i j2 ...jn + · · · + Γlk Aj1 j2 ...jn 2 ...jn 5.4 Exercises and Revision 135 i1 i2 ...im m 2 ...im − · · · − Γljn k Aji11ij22...i − Γlj2 k Aij11il...j −Γlj1 k Alj ...l n 2 ...jn 5.35 5.36 5.37 5.38 5.39 5.40 5.41 5.42 5.43 5.44 5.45 5.46 5.47 5.48 5.49 5.50 5.51 5.52 5.53 Extract from the pattern of this expression the practical rules that should be followed in writing the analytical expressions of covariant derivative of tensors of any rank and type. In the expression of covariant derivative, what the partial derivative term stands for and what the Christoffel symbol terms represent? For the covariant derivative of a type (m, n, w) tensor, obtain the number of total terms, the number of negative Christoffel symbol terms and the number of positive Christoffel symbol terms. What is the rank and type of the covariant derivative of a tensor of rank-n and type (p, q)? The covariant derivative of a differentiable scalar function is the same as the ordinary partial derivative. Why? What is the significance of the dependence of the covariant derivative on the Christoffel symbols with regard to its relation to the space and coordinate system? The covariant derivative of tensors in coordinate systems with constant basis vectors is the same as the ordinary partial derivative for all tensor ranks. Why? Express, mathematically, the fact that the metric tensor is in lieu of constant with respect to covariant differentiation. Which rules of ordinary partial differentiation also apply to covariant differentiation and which rules do not? State all these rules symbolically for both ordinary and covariant differentiation. Explain why the covariant differential operators with respect to different indices do not commute, i.e. ∂;i ∂;j 6= ∂;j ∂;i (i 6= j). State the Ricci theorem about covariant differentiation of the metric tensor and prove it with full justification of each step. State, symbolically, the commutative property of the covariant derivative operator with the index shifting operator (which is based on the Ricci theorem) using the symbolic notation one time and the indicial notation another. Verify that the ordinary Kronecker delta tensor is constant with respect to covariant differentiation. State, symbolically, the fact that covariant differentiation and contraction of index operations commute with each other. What is the condition on the components of the metric tensor that makes the covariant derivative become ordinary partial derivative for all tensor ranks? Prove that covariant differentiation and contraction of indices commute. What is the mathematical condition that is required if the mixed second order partial derivatives should be equal, i.e. ∂i ∂j = ∂j ∂i (i 6= j)? What is the mathematical condition that is required if the mixed second order covariant derivatives should be equal, i.e. ∂;i ∂;j = ∂;j ∂;i (i 6= j)? Derive analytical expressions for Ai;jk and Ai;kj and hence verify that Ai;jk 6= Ai;kj . From the result of exercise 5.52 plus Eq. 560, verify the following relation: Ai;jk − 5.4 Exercises and Revision 5.54 5.55 5.56 5.57 5.58 136 Ai;kj = Aa Raijk . What is the covariant derivative of a relative scalar f of weight w? What is the covariant derivative of a rank-2 relative tensor Aij of weight w? Why the covariant derivative of a non-scalar tensor with constant components is not necessarily zero in general coordinate systems? Which term of the covariant derivative of such a tensor will vanish? Show that: Ai;j = Aj;i where A is a gradient of a scalar field. Show that the covariant derivative of the basis vectors of the covariant and contravariant types is identically zero, i.e. Ei;j = 0 and Ei;j = 0. Prove the following identity:  ∂k gij Ai B j = Ai;k B i + Ai Bi;k 5.59 Define absolute differentiation descriptively and mathematically. What are the other names of absolute derivative? 5.60 Write the mathematical expression for the absolute derivative of the tensor field Aijk which is defined over a space curve C(t). 5.61 Why the absolute derivative of a differentiable scalar is the same as its ordinary total derivative, i.e. δf = df ? δt dt 5.62 Why the absolute derivative of a differentiable non-scalar tensor is the same as its ordinary total derivative in rectilinear coordinate systems? 5.63 From the pattern of covariant derivative of a general tensor, obtain the pattern of its absolute derivative. ) 5.64 We have A = Aijk Ei Ej Ek . Apply the ordinary total differentiation process (i.e. dA dt onto this tensor (including its basis vectors) to obtain its absolute derivative. 5.65 Which rules of ordinary total differentiation also apply to intrinsic differentiation and which rules do not? State all these rules symbolically for both ordinary and intrinsic differentiation. 5.66 Using your knowledge about covariant differentiation and the fact that absolute differentiation follows the style of covariant differentiation, obtain all the rules of absolute differentiation of the metric tensor, the Kronecker delta tensor and the index shifting and index replacement operators. Express all these rules in words and in symbols. 5.67 Justify the following statement: “For coordinate systems in which all the components of the metric tensor are constants, the absolute derivative is the same as the ordinary total derivative”. 5.68 The absolute derivative of a tensor along a given curve is unique. What this means? 5.69 Summarize all the main properties and rules that govern covariant and absolute differentiation. Chapter 6 Differential Operations In this chapter, we examine the main differential operations which are based on the nabla operator ∇ as defined in tensor calculus using largely tensor notation. These operations are based on the various types of interaction between the vector differential operator nabla and tensors of different ranks where some of these interactions involve the dot and cross product operations. The chapter will investigate these operations in general coordinate systems and in general orthogonal coordinate systems which are a special case of the general coordinate systems. The chapter will also investigate these operations in Cartesian coordinate systems as well as the two most important and widely used curvilinear orthogonal coordinate systems, namely the cylindrical and spherical systems, because of their particular importance and widespread application in science and engineering. Due to the likely familiarity of the reader with these operations in Cartesian systems, which is usually acquired at this level from a previous course on vector calculus, we start our investigation from the Cartesian system. This may help to remind the reader and make the subsequent sections easier to understand. Regarding the cylindrical and spherical coordinate systems, we can use indexed general coordinates like u1 , u2 and u3 to represent the cylindrical coordinates (ρ, φ, z) and the spherical coordinates (r, θ, φ) and hence we can express these operations in tensor notation as we do for the other systems. However, for the sake of clarity and to follow the more conventional practice, we use the coordinates of these systems as suffixes in place of the usual indices used in the tensor notation. In this context, we should insist that these suffixes are labels and not indices and therefore they do not follow the rules and conventions of tensor indices such as following the summation convention. In fact, there is another reason for the use of suffix labels instead of symbolic indices that is the components in the cylindrical and spherical coordinates are physical, not covariant or contravariant, and hence suffixing with coordinates is more appropriate (see § 3.3). Before we start this investigation, we should remark that the differentiation of a tensor increases its rank by one, by introducing an extra covariant index, unless it implies a contraction in which case it reduces the rank by one. Therefore the gradient of a scalar is a vector and the gradient of a vector is a rank-2 tensor, while the divergence of a vector is a scalar and the divergence of a rank-2 tensor is a vector. This may be justified by the fact that the gradient operator is a vector operator. On the other hand the Laplacian operator does not change the rank since it is a scalar operator; hence the Laplacian of a scalar is a scalar and the Laplacian of a vector is a vector and so on. We should also remark that there are other nabla based operators and operations which are subsidiary to the main ones and are used in pure and applied mathematics and science. For example, the following scalar differential operator, defined in Cartesian coordinates, 137 6.1 Cartesian Coordinate System 138 is commonly used in science such as fluid dynamics: A · ∇ = Ai ∇i = Ai ∂ = Ai ∂i ∂xi (413) where A is a vector. As explained earlier (see § 1.2), the order of Ai and ∂i should be respected. Another example is the following vector differential operator which also has common applications in science: [A × ∇]i = ǫijk Aj ∂k (414) where, again, the order should be respected and the operator is defined in Cartesian coordinates. 6.1 6.1.1 Cartesian Coordinate System Operators The nabla vector operator ∇ is a spatial partial differential operator which is defined in Cartesian coordinate systems by: ∂ ∂xi Similarly, the Laplacian scalar operator is given by: ∇i = ∇2 = δij ∂2 ∂2 = = ∇ii = ∂ii ∂xi ∂xj ∂xi ∂xi (415) (416) We note that the Laplacian operator may also be notated with ∆ (as well as several other symbols such as ∂ii2 ). However, in the present book we do not use this notation for this purpose. 6.1.2 Gradient Based on the above definition of nabla, the gradient of a differentiable scalar function of position f is a vector obtained by applying the nabla operator to f and hence it is defined by: ∂f = ∂i f = f,i (417) [∇f ]i = ∇i f = ∂xi Similarly, the gradient of a differentiable vector field A is the outer product (refer to § 3.2.3) between the nabla operator and the vector and hence it is a rank-2 tensor, that is: [∇A]ij = ∂i Aj This definition can be easily extended to higher rank tensors. (418) 6.1.3 Divergence 6.1.3 139 Divergence The divergence of a differentiable vector A is the dot product of the nabla operator and the vector A and hence it is a scalar given by: ∇ · A = δij ∂Ai ∂Ai = = ∇i Ai = ∂i Ai = Ai,i ∂xj ∂xi (419) The divergence operation can also be viewed as taking the gradient of the vector followed by a contraction. Hence, the divergence of a vector is invariant because it is the trace of a rank-2 tensor (see § 7.1.2). It may also be argued more simply that the divergence of a vector is a scalar and hence it is invariant. Similarly, the divergence of a differentiable rank-2 tensor A is a vector defined in one of its forms by: [∇ · A]i = ∂j Aji (420) and in another form by: (421) [∇ · A]j = ∂i Aji These two forms may be given respectively, using the symbolic notation, by: ∇·A and ∇ · AT (422) where AT is the transpose of A. More generally, the divergence of a tensor of rank n ≥ 2, which is a tensor of rank-(n − 1), can be defined in several forms, which are different in general, depending on the choice of the contracted index. 6.1.4 Curl The curl of a differentiable vector A is the cross product of the nabla operator and the vector A and hence it is a vector defined by: [∇ × A]i = ǫijk ∂Ak = ǫijk ∇j Ak = ǫijk ∂j Ak = ǫijk Ak,j ∂xj (423) The curl operation may be generalized to tensors of rank > 1 (see for example § 7.1.4), and hence the curl of a differentiable rank-2 tensor A can be defined as a rank-2 tensor given by: [∇ × A]ij = ǫimn ∂m Anj (424) The last example can be easily extended to higher rank tensors. We note that there is more than one possibility for the contraction of the last index of the permutation tensor with one of the tensor indices when the rank is > 1. 6.1.5 Laplacian 6.1.5 140 Laplacian The Laplacian of a differentiable scalar f is given by: ∂ 2f ∂ 2f = = ∇ii f = ∂ii f = f,ii ∇ f = δij ∂xi ∂xj ∂xi ∂xi 2 (425) The Laplacian of a differentiable vector A is defined for each component of the vector in a similar manner to the definition of the Laplacian acting on a scalar, that is:  2  (426) ∇ A i = ∇2 [A]i = ∂jj Ai This definition can be easily extended to higher rank tensors. 6.2 General Coordinate System Here, we investigate the differential operators and operations in general coordinate systems. We note that the definitions of the differential operations in Cartesian systems, as given in the vector calculus texts and as outlined in § 6.1, are essentially valid in general non-Cartesian coordinate systems if the operations are extended to include the basis vectors as well as the components, as we will see in the following subsections. We also note that the analytical expressions of the differential operations can be obtained directly if the expression for the nabla operator ∇ and the spatial derivatives of the basis vectors of the given coordinate system are known. 6.2.1 Operators The nabla operator ∇, which is a spatial partial differential vector operator, is defined in general coordinate systems by: ∇ = Ei ∂ i (427) Similarly, the Laplacian operator is defined generically by: ∇2 = div grad = ∇ · ∇ (428) More details about these operators will follow in the next subsections. 6.2.2 Gradient Based on the definition of the nabla operator, as given in the previous subsection, the gradient of a differentiable scalar function of position, f , is given by: ∇f = Ei ∂i f = Ei f,i (429) The components of this expression represent the covariant form of a rank-1 tensor, i.e. [∇f ]i = f,i , as it should be since the gradient operation increases the covariant rank of a tensor by one, as indicated earlier. Since this expression consists of a contravariant basis 6.2.2 Gradient 141 vector and a covariant component, the gradient in general coordinate systems is invariant under admissible coordinate transformations. The gradient of a differentiable covariant vector A can similarly be defined as follows:[27]  (430) ∇A = Ei ∂i Aj Ej = Ei Ej ∂i Aj + Ei Aj ∂i Ej (product rule) = Ei Ej ∂i Aj + Ei Aj −Γjki E = Ei Ej ∂i Aj − Ei Ej Γkji Ak  = Ei Ej ∂i Aj − Γkji Ak  k (Eq. 313) (relabeling dummy indices j & k) (taking common factor Ei Ej ) = Ei Ej Aj;i (definition of covariant derivative) and hence it is the covariant derivative of the vector. Similarly, for a differentiable contravariant vector A the gradient is given by:  (431) ∇A = Ei ∂i Aj Ej = Ei Ej ∂i Aj + Ei Aj ∂i Ej = Ei Ej ∂i Aj + Ei Aj Γkji Ek = = = Ei Ej ∂i Aj + Ei Ej Γjki Ak  Ei Ej ∂i Aj + Γjki Ak Ei Ej Aj;i (product rule)  (Eq. 312) (relabeling dummy indices j & k) (taking common factor Ei Ej ) (definition of covariant derivative) The components of the gradients of covariant and contravariant vectors represent, respectively, the covariant and mixed forms of a rank-2 tensor, as they should be since the gradient operation increases the covariant rank of a tensor by one. The gradient of higher rank tensors is similarly defined. For example, the gradient of a rank-2 tensor in its covariant, contravariant and mixed form is given by:  (covariant) (432) ∇A = Ei Ej Ek ∂i Ajk − Γlji Alk − Γlki Ajl = Ei Ej Ek Ajk;i  jk j ∇A = Ei Ej Ek ∂i Ajk + Γli Alk + Γkli Ajl = Ei Ej Ek A ;i (contravariant) (433)  i j k k l k k l i j (mixed) (434) ∇A = E E Ek ∂i Aj − Γji Al + Γli Aj = E E Ek Aj;i  j j l j i k j l i k (mixed) (435) ∇A = E Ej E ∂i A k + Γli A k − Γki A l = E Ej E A k;i We finally remark that the contravariant form of the gradient operation can be obtained by using the index raising operator. For example, the contravariant form of the gradient of a scalar f is given by: [∇f ]i = ∂ i f = g ij ∂j f = g ij f,j = f ,i [27] (436) We note that the basis vector Ei which associates the derivative operator in the following equations (as well as in similar equations and expressions) should be the last one in the basis tensor so that the order of the indices inthe components and in the basis tensor are the same. So, strictly we should write ∇A = ∂i Aj Ej Ei where ∂i acts only on what is inside the parentheses and hence the final expression becomes ∇A = Ej Ei Aj;i . However, to avoid confusion we put the vector to the left relying on this understanding. 6.2.3 Divergence 142 This can be easily extended to higher rank tensors. As seen, the contravariant form of the gradient operation increases the contravariant rank of the tensor by one. 6.2.3 Divergence Generically, the divergence of a differentiable contravariant vector A is defined as follows:  ∇ · A = E i ∂ i · Aj E j (437)  = Ei · ∂i Aj Ej  (definition of covariant derivative) = Ei · Aj;i Ej  j = Ei · Ej A ;i = δ ij Aj;i (Eq. 215) Ai;i (Eq. 172) = In more details, the divergence of a differentiable contravariant vector Ai is a scalar obtained by contracting the covariant derivative index with the contravariant index of the vector, and hence: ∇ · A = Ai;i (438) i = ∂i A + Γiji Aj 1 √ = ∂i Ai + Aj √ ∂j ( g) g 1 √ = ∂i Ai + Ai √ ∂i ( g) g 1 √ i = √ ∂i gA g (Eq. 362) (Eq. 324) (renaming dummy index j) (product rule) where g is the determinant of the covariant metric tensor gij . The last equality may be called the Voss-Weyl formula. The divergence can also be obtained by raising the first index of the covariant derivative of a covariant vector using a contracting contravariant metric tensor, that is:  (Eq. 376) (439) g ji Aj;i = g ji Aj ;i  (Eq. 228) = Ai ;i = Ai;i =∇·A (Eq. 438) as before. Accordingly, the divergence of a covariant vector Aj is obtained by using the raising operator, that is: Ai;i = g ij Aj;i (440) 6.2.4 Curl 143 For a rank-2 contravariant tensor A, the divergence is generically defined by:  ∇ · A = Ei ∂i · Ajk Ej Ek (441)  = Ei · ∂i Ajk Ej Ek   jk i (definition of covariant derivative) = E · A ;i Ej Ek  = Ei · Ej Ek Ajk;i = δji Ek Ajk;i (Eq. 215) = Ek Aik;i (Eq. 172) The components of this expression represent a contravariant vector, as it should be since the divergence operation reduces the contravariant rank of a tensor by one. More generally, considering the tensor components, the divergence of a differentiable rank-2 contravariant tensor Aij is a contravariant vector obtained by contracting the covariant derivative index with one of the contravariant indices, that is: [∇ · A]j = Aij;i or [∇ · A]i = Aij;j (442) And for a rank-2 mixed tensor Aij we have: [∇ · A]j = Aij;i (443) k ···im Similarly, for a general tensor of type (m, n): A = Aji11ij22···i ·········jn , the divergence with respect to its k th contravariant index is defined by:  ···im i1 i2 ···s···im = A [∇ · A]ji11ij22············j (444) j1 j2 ·········jn ;s n with the absence of the contracted contravariant index ik on the left hand side. As a matter of notation, it should be understood that ∇ · A is lower than the original tensor A by just one contravariant index and hence, unlike the common use of this notation, it is not necessarily scalar. 6.2.4 Curl The curl of a differentiable vector is the cross product of the nabla operator ∇ with the vector. For example, the curl of a vector A represented by covariant components is given by: (445) curl A = ∇ × A = Ei ∂i × Aj Ej  = Ei × ∂i Aj Ej  = Ei × Aj;i Ej  = Aj;i Ei × Ej (Eq. 359) 6.2.5 Laplacian 144 = Aj;i ǫijk Ek (Eq. 274) ijk = ǫ Aj;i Ek  ǫijk = √ ∂i Aj − Γlji Al Ek g (Eqs. 166 & 361) Hence, the k th contravariant component of curl A is given by:  ǫijk [∇ × A]k = √ ∂i Aj − Γlji Al g (446) On expanding the last equation for the three components of a vector in a 3D space, considering that the terms of the Christoffel symbols cancel out due to their symmetry in the two lower indices,[28] we obtain: 1 [∇ × A]1 = √ (∂2 A3 − ∂3 A2 ) g 1 [∇ × A]2 = √ (∂3 A1 − ∂1 A3 ) g 1 [∇ × A]3 = √ (∂1 A2 − ∂2 A1 ) g (447) (448) (449) Hence, Eq. 446 will reduce to: ǫijk [∇ × A]k = √ ∂i Aj g 6.2.5 (450) Laplacian Generically, the Laplacian of a differentiable scalar function of position, f , is defined as follows: ∇2 f = div (grad f ) = ∇ · (∇f ) (451) Hence the simplest approach for obtaining the Laplacian in general coordinate systems is to insert the expression for the gradient, ∇f , into the expression for the divergence. However, because in general coordinate systems the divergence is defined only for tensors having at least one contravariant index to facilitate the contraction with the covariant [28] That is: Ai;j − Aj;i = = = = ∂j Ai − Ak Γkij − ∂i Aj + Ak Γkji ∂j Ai − Ak Γkij − ∂i Aj + Ak Γkij ∂j Ai − ∂i Aj Ai,j − Aj,i 6.2.5 Laplacian 145 derivative index (see for example Eqs. 437 and 444) whereas the gradient of a scalar is a covariant tensor, the index of the gradient should be raised first before applying the divergence operation, that is: [∇f ]i = ∂ i f = g ij ∂j f (452) Now, according to Eq. 438 the divergence is given by: 1 √ i ∇ · A = Ai;i = √ ∂i gA g (453) On defining A ≡ Ei ∂ i f and replacing Ai in Eq. 453 with ∂ i f using Eq. 452 we obtain:   1 √ ij ∇2 f = ∇ · Ei ∂ i f = √ ∂ i gg ∂j f g (454) which is the expression for the Laplacian of a scalar function f in general coordinate systems. Another approach for developing the Laplacian expression in general coordinate systems is to apply the first principles by using the definitions and basic properties of the operations involved, that is: ∇2 f = ∇ · (∇f ) i (455) j = E ∂i · E ∂j f   = Ei · ∂ i Ej ∂ j f  = Ei · ∂i Ej f,j  = Ei · Ej f,j;i  = Ei · Ej f,j;i (Eq. 359) = g ij f,j;i ij = g f,j  = g ij ∂j f ij (Eq. 214) ;i (Eq. 376)  (Eq. 362) ;i   = ∂i g ∂j f + g kj ∂j f Γiki   = ∂i g ij ∂j f + g ij ∂j f Γkik   1 √ = ∂i g ij ∂j f + g ij ∂j f √ (∂i g) g   √  1 √ =√ g∂i g ij ∂j f + g ij ∂j f ∂i g g  1 √ ij = √ ∂i gg ∂j f g (renaming dummy indices i & k) (Eq. 324) 1 (taking √ factor) g (product rule) which is the same as before (see Eq. 454). The Laplacian of a scalar f may also be shorthand notated with: ∇2 f = g ij f,ij (456) 6.3 Orthogonal Coordinate System 146 The Laplacian of non-scalar tensors can be similarly defined. For example, the Laplacian of a vector B in its contravariant and covariant forms, B i and Bi , is a vector A (i.e. A = ∇2 B) which may be defined in general coordinate systems as: Ai = g jk B i ;jk Ai = g jk Bi;jk (457) As indicated earlier, the Laplacian of a tensor is a tensor of the same rank and variance type. 6.3 Orthogonal Coordinate System In this section, we state the main differential operators and operations in general orthogonal coordinate systems. These operators and operations are special cases of those of general coordinate systems which were derived in § 6.2. However, due to the widespread use of orthogonal coordinate systems, it is beneficial to state the most important of these operators and operations although they can be easily obtained from the formulae of general coordinate systems. For clarity, general orthogonal coordinate systems are identified in this section by the coordinates (q 1 , . . . , q n ) with unit basis vectors (q1 , . . . , qn ) and scale factors (h1 , . . . , hn ) where: X 1 ∂xj X ∂q i qi = e = h i j ej (no sum on i) (458) j hi ∂q i ∂x j j #1/2 " #−1/2 " X  ∂q i 2 X  ∂xj 2 ∂r = hi = |Ei | = = ∂q i ∂q i ∂xj j j (459) In the last equations, xj and ej are respectively the coordinates and unit basis vectors in a Cartesian rectangular system, and r = xi ei is the position vector in that system. We remark that in orthogonal coordinate systems the covariant and contravariant normalized basis vectors are identical, as established previously in § 3.3, and hence qi = qi and ej = ej . 6.3.1 Operators The nabla operator in general orthogonal coordinate systems is given by: ∇= X qi ∂ hi ∂q i i (460) while the Laplacian operator, assuming a 3D space, is given by:   3 1 X ∂ h1 h2 h3 ∂ ∇ = h1 h2 h3 i=1 ∂q i (hi )2 ∂q i 2 (461) 6.3.2 Gradient 6.3.2 147 Gradient The gradient of a differentiable scalar f in orthogonal coordinate systems, assuming a 3D space, is given by: 3 X q1 ∂f q2 ∂f q3 ∂f qi ∂f = + + ∇f = i 1 2 h ∂q h1 ∂q h2 ∂q h3 ∂q 3 i=1 i 6.3.3 (462) Divergence The divergence in orthogonal coordinate systems can be obtained from Eq. 438. Since for orthogonal coordinate systems the metric tensor according to Eqs. 233-236 is diagonal √ with g = h1 h2 h3 in a 3D space and hi Ai = Âi (no summation) according to Eq. 144, the last line of Eq. 438 becomes: 1 ∂ √ i gA ∇·A = √ g ∂q i   3 1 X ∂ h1 h2 h3 i = Â h1 h2 h3 i=1 ∂q i hi     ∂  ∂  1 ∂  h2 h3 Â1 + 2 h1 h3 Â2 + 3 h1 h2 Â3 = h1 h2 h3 ∂q 1 ∂q ∂q (463) where A is a contravariant differentiable vector and Âi represents its physical components. This equation is the divergence of a vector in general orthogonal coordinates as defined in vector calculus. We note that in orthogonal coordinate systems the physical components are the same for covariant and contravariant forms, as established before in § 3.3, and hence Âi = Âi . 6.3.4 Curl The curl of a differentiable vector A in orthogonal coordinate systems in 3D spaces is given by: h1 q 1 h2 q2 h3 q3 1 ∂ ∂ ∂ ∇×A= (464) ∂q 1 ∂q 2 ∂q 3 h1 h2 h3 h1 Â1 h2 Â2 h3 Â3 where the hat indicates a physical component. The last equation may also be given in a more compact form as: 3 X ǫijk hi ∂(hk Âk ) [∇ × A]i = h1 h2 h3 ∂q j k=1 (no sum on i) (465) 6.3.5 Laplacian 6.3.5 148 Laplacian As seen earlier, for general orthogonal coordinate systems in 3D spaces we have: √ g ii = g = h1 h2 h3 1 (hi )2 (no sum) g ij = 0 (i 6= j) (466) and hence Eq. 454 becomes:   3 1 X ∂ h1 h2 h3 ∂f ∇f= h1 h2 h3 i=1 ∂q i (hi )2 ∂q i 2 (467) which is the Laplacian of a scalar function of position, f , in orthogonal coordinate systems as defined in vector calculus. 6.4 Cylindrical Coordinate System For cylindrical coordinate systems identified by the coordinates (ρ, φ, z), the orthonormal basis vectors are eρ , eφ and ez . Although the given components (i.e. Aρ , Aφ and Az ) are physical components, we do not use hats since the components are suffixed with coordinate symbols (refer to § 3.3). We use for brevity eρφ as a shorthand notation for the unit dyad eρ eφ and similar notations for the other dyads. We remark that the following expressions for the operators and operations in cylindrical coordinate systems can be obtained from the definition of these operators and operations in general coordinate systems using the metric tensor of the cylindrical system (see § 4.5). They can also be obtained more simply from the corresponding expressions in orthogonal coordinate systems using the scale factors of the cylindrical system in Table 1. It should be obvious that since ρ, φ and z are labels for specific coordinates and not variable indices, the summation convention does not apply to these labels. We note that cylindrical coordinate systems are defined specifically for 3D spaces. We also note that the following operators and operations can be obtained for the 2D plane polar coordinate systems by dropping the z components or terms from the cylindrical form of these operators and operations. 6.4.1 Operators The nabla operator ∇ in cylindrical coordinate systems is given by: 1 ∇ = eρ ∂ ρ + eφ ∂ φ + ez ∂ z ρ (468) while the Laplacian operator is given by: 1 1 ∇2 = ∂ρρ + ∂ρ + 2 ∂φφ + ∂zz ρ ρ where ∂ρρ = ∂ρ ∂ρ and the notation equally applies to other similar symbols. (469) 6.4.2 Gradient 6.4.2 149 Gradient The gradient of a differentiable scalar f is given by: 1 ∇f = eρ ∂ρ f + eφ ∂φ f + ez ∂z f ρ (470) The gradient of a differentiable vector A is given by: ∇A = eρρ Aρ,ρ + eρφ Aφ,ρ + eρz Az,ρ +     1 1 1 Aφ Aρ + eφφ + eφz Az,φ + Aρ,φ − Aφ,φ + eφρ ρ ρ ρ ρ ρ ezρ Aρ,z + ezφ Aφ,z + ezz Az,z (471) where eρρ , eρφ , · · · , ezz are unit dyads as defined above. 6.4.3 Divergence The divergence of a differentiable vector A is given by: ∇·A= 1 [∂ρ (ρAρ ) + ∂φ Aφ + ρ∂z Az ] ρ The divergence of a differentiable rank-2 tensor A is given by:   Aρρ − Aφφ 1 + Aφρ,φ + Azρ,z + ∇ · A = eρ Aρρ,ρ + ρ ρ   2Aρφ 1 Aφρ − Aρφ eφ Aρφ,ρ + + Aφφ,φ + Azφ,z + + ρ ρ ρ   Aρz 1 ez Aρz,ρ + + Aφz,φ + Azz,z ρ ρ (472) (473) We note that Eq. 472 can be obtained by contracting Eq. 471. 6.4.4 Curl The curl of a differentiable vector A is given by: ∇×A= 6.4.5 1 ρ eρ ρeφ ez ∂ρ ∂φ ∂z Aρ ρAφ Az (474) Laplacian The Laplacian of a differentiable scalar f is given by: 1 1 ∇2 f = ∂ρρ f + ∂ρ f + 2 ∂φφ f + ∂zz f ρ ρ (475) 6.5 Spherical Coordinate System The Laplacian of a differentiable vector A is given by:     2 1 1 2 ∂ρ (ρAρ ) + 2 ∂φφ Aρ + ∂zz Aρ − 2 ∂φ Aφ + ∇ A = eρ ∂ ρ ρ ρ ρ     1 2 1 eφ ∂ ρ ∂ρ (ρAφ ) + 2 ∂φφ Aφ + ∂zz Aφ + 2 ∂φ Aρ + ρ ρ ρ   1 1 ez ∂ρ (ρ∂ρ Az ) + 2 ∂φφ Az + ∂zz Az ρ ρ 6.5 150 (476) Spherical Coordinate System For spherical coordinate systems identified by the coordinates (r, θ, φ), the orthonormal basis vectors are er , eθ and eφ . As in the case of cylindrical coordinates, the components are physical and we do not use hats for the same reason. We use for brevity erθ as a shorthand notation for the dyad er eθ and similar notations for the other unit dyads. As for cylindrical systems, the following expressions for the operators and operations in spherical coordinate systems can be obtained from the corresponding definitions in general coordinate systems using the metric tensor of the spherical system (see § 4.5) or from the corresponding expressions in orthogonal systems using the scale factors of the spherical system in Table 1. Again, the summation convention does not apply to r, θ and φ since they are labels and not indices. We also note that spherical coordinate systems are defined specifically for 3D spaces. 6.5.1 Operators The nabla operator ∇ in spherical coordinate systems is given by: 1 1 ∂φ ∇ = er ∂ r + eθ ∂ θ + eφ r r sin θ (477) while the Laplacian operator is given by: cos θ 1 1 2 ∂θ + 2 2 ∂φφ ∇2 = ∂rr + ∂r + 2 ∂θθ + 2 r r r sin θ r sin θ (478) where ∂rr = ∂r ∂r and the notation equally applies to other similar symbols. 6.5.2 Gradient The gradient of a differentiable scalar f in spherical coordinate systems is given by: 1 1 ∇f = er ∂r f + eθ ∂θ f + eφ ∂φ f r r sin θ (479) The gradient of a differentiable vector A is given by: ∇A = err Ar,r + erθ Aθ,r + erφ Aφ,r + (480) 6.5.3 Divergence 151    Ar,θ Aθ Aθ,θ Ar Aφ,θ − + eθθ + + eθφ + eθr r r r r r       Aθ,φ Aφ,φ Aφ Aφ cot θ Ar Aθ cot θ Ar,φ + eφθ + eφφ − − + + eφr r sin θ r r sin θ r r sin θ r r  where err , erθ , · · · , eφφ are unit dyads as defined above. 6.5.3 Divergence The divergence of a differentiable vector A is given by:   ∂ (r2 Ar ) ∂ (sin θAθ ) ∂Aφ 1 sin θ +r +r ∇·A= 2 r sin θ ∂r ∂θ ∂φ (481) The divergence of a differentiable rank-2 tensor A is given by:   ∂r (r2 Arr ) ∂θ (Aθr sin θ) ∂φ Aφr Aθθ + Aφφ ∇ · A = er + + − + (482) r2 r sin θ r sin θ r   ∂r (r3 Arθ ) ∂θ (Aθθ sin θ) ∂φ Aφθ Aθr − Arθ − Aφφ cot θ eθ + + + + r3 r sin θ r sin θ r   ∂r (r3 Arφ ) ∂θ (Aθφ sin θ) ∂φ Aφφ Aφr − Arφ + Aφθ cot θ + eφ + + r3 r sin θ r sin θ r We note that Eq. 481 can be obtained by contracting Eq. 480. 6.5.4 Curl The curl of a differentiable vector A is given by: ∇×A= 6.5.5 1 r2 sin θ er reθ r sin θeφ ∂r ∂θ ∂φ Ar rAθ r sin θAφ (483) Laplacian The Laplacian of a differentiable scalar f is given by: 2 cos θ 1 1 ∇2 f = ∂rr f + ∂r f + 2 ∂θθ f + 2 ∂θ f + 2 2 ∂φφ f (484) r r r sin θ r sin θ The Laplacian of a differentiable vector A is given by:     2∂θ (Aθ sin θ) ∂θ (sin θ∂θ Ar ) ∂φφ Ar 2∂φ Aφ ∂r (r2 Ar ) 2 + + 2 2 − − 2 +(485) ∇ A = er ∂ r r2 r2 sin θ r2 sin θ r sin θ r sin θ     ∂r (r2 ∂r Aθ ) 1 ∂θ (Aθ sin θ) 2 cot θ ∂φφ Aθ 2∂θ Ar eθ + 2 2 + + 2 ∂θ − 2 ∂ φ Aφ + r2 r sin θ r2 r sin θ r sin θ     ∂φφ Aφ 1 ∂θ (Aφ sin θ) 2∂φ Ar 2 cot θ ∂r (r2 ∂r Aφ ) + 2 2 + 2 + 2 ∂θ + ∂ φ Aθ eφ r2 r sin θ r sin θ r sin θ r2 sin θ 6.6 Exercises and Revision 6.6 152 Exercises and Revision 6.1 Describe briefly the nabla based differential operators and operations considering the interaction of the nabla operator with the tensors which are acted upon by this operator. 6.2 What are the advantages and disadvantages of using the coordinates as suffixes for labeling the operators, basis vectors and tensor components in cylindrical and spherical systems instead of indexed general coordinates? What are the advantages and disadvantages of the opposite? 6.3 “The differentiation of a tensor increases its rank by one, by introducing an extra covariant index, unless it implies a contraction in which case it reduces the rank by one”. Justify this statement giving common examples from vector and tensor calculus. 6.4 Write the following subsidiary nabla based operators in tensor notation: A · ∇ and A × ∇. Is this notation consistent with the notation of dot and cross product of vectors? 6.5 Why in general we have: A · ∇ = 6 ∇ · A and A × ∇ 6= ∇ × A? 6.6 Define the nabla vector operator and the Laplacian scalar operator in Cartesian coordinate systems using tensor notation. 6.7 Find the gradient of the following vector field in a Cartesian coordinate system: A = (x, 2x2 , π). 6.8 Define the divergence of a differentiable vector descriptively and mathematically assuming a Cartesian coordinate system. 6.9 What is the divergence of the following vector field in Cartesian coordinates: A = (2z, y 3 , ex )? 6.10 Write symbolically, using tensor notation, the following two forms of the divergence of a rank-2 tensor field A in Cartesian coordinates: ∇ · A and ∇ · AT . 6.11 Define the curl ∇ × A in Cartesian coordinates using tensor notation where (a) A is a rank-1 tensor and (b) A is a rank-2 tensor (note the two possibilities in the last case). 6.12 What is the curl of the following vector field assuming a Cartesian coordinate system: A = (5e2x , πxy, z 2 )? 6.13 Find the Laplacian of the following vector field in Cartesian coordinates: A = (x2 y, 2y sin z, πzecosh x ). 6.14 Define the nabla operator and the Laplacian operator in general coordinate systems using tensor notation. 6.15 Obtain an expression for the gradient of a covariant vector A = Ai Ei in general coordinates justifying each step in your derivation. Repeat the question with a contravariant vector A = Ai Ei . 6.16 Repeat question 6.15 with a rank-2 mixed tensor A = Aij Ei Ej . 6.17 Define, in tensor language, the contravariant form of the gradient of a scalar field. 6.18 Define the divergence of a differentiable vector descriptively and mathematically assuming a general coordinate system. 6.19 Derive the following expression for the divergence of a contravariant vector A in 6.6 Exercises and Revision 6.20 6.21 6.22 6.23 153 general coordinates: ∇ · A = Ai;i . Verify the following formula for the divergence of a contravariant vector A in general √ i coordinates: ∇ · A = √1g ∂i gA . Repeat the question with the formula: ∇ · A = g ji Aj;i where A is a covariant vector. Repeat question 6.20 with the formula: ∇ · A = Ek Aik;i where A is a rank-2 contravariant tensor. Prove that the divergence of a contravariant vector is a scalar (i.e. rank-0 tensor) by showing that it is invariant under coordinate transformations. Derive, from the first principles, the following formula for the curl  of a covariant vector ijk field A in general coordinates: [∇ × A]k = ǫ√g ∂i Aj − Γlji Al . 6.24 Show that the formula in exercise 6.23 will reduce to [∇ × A]k = ǫ√g ∂i Aj due to the symmetry of the Christoffel symbols in their lower indices. 6.25 Derive, from the first principles, the following expression  for the Laplacian of a scalar √ ij field f in general coordinates: ∇2 f = √1g ∂i gg ∂j f . 6.26 Why the basic definition of the Laplacian of a scalar field f in general coordinates as ∇2 f = div (grad f ) cannot be used as it is to develop a formula before raising the index of the gradient? 6.27 Define, in tensor language, the nabla operator and the Laplacian operator assuming an orthogonal coordinate system of a 3D space. 6.28 Using the expression of the divergence of a vector field in general coordinates, obtain an expression for the divergence in orthogonal coordinates. 6.29 Define the curl of a vector field A in orthogonal coordinates of a 3D space using determinantal form and tensor notation form. 6.30 Using the expression of the Laplacian of a scalar field in general coordinates, derive an expression for the Laplacian in orthogonal coordinates of a 3D space. 6.31 Why the components of tensors in cylindrical and spherical coordinates are physical? 6.32 Define the nabla and Laplacian operators in cylindrical coordinates. 6.33 Use the definition of the gradient of a scalar field f in orthogonal coordinates and the scale factors of Table 1 to obtain an expression for the gradient in cylindrical coordinates. 6.34 Use the definition of the divergence of a vector field A in orthogonal coordinates and the scale factors of Table 1 to obtain an expression for the divergence in cylindrical coordinates. 6.35 Write the determinantal form of the curl of a vector field A in cylindrical coordinates. 6.36 Use the definition of the Laplacian of a scalar field f in orthogonal coordinates and the scale factors of Table 1 to obtain an expression for the Laplacian in cylindrical coordinates. 6.37 A scalar field in cylindrical coordinates is given by: f (ρ, φ, z) = ρ. What are the gradient and Laplacian of this field? 6.38 A vector field in cylindrical coordinates is given by: A (ρ, φ, z) = (3z, πφ2 , z 2 cos ρ). What are the divergence and curl of this field? 6.39 Repeat exercise 6.33 with spherical coordinates. ijk 6.6 Exercises and Revision 6.40 6.41 6.42 6.43 154 Repeat exercise 6.34 with spherical coordinates. Repeat exercise 6.35 with spherical coordinates. Repeat exercise 6.36 with spherical coordinates. A scalar field in spherical coordinates is given by: f (r, θ, φ) = r2 + θ. What are the gradient and Laplacian of this field? 6.44 A vector field in spherical coordinates is given by: A (r, θ, φ) = (er , 5 sin φ, ln θ). What are the divergence and curl of this field? Chapter 7 Tensors in Application In this chapter, we conduct a preliminary investigation about some tensors and tensor notation and techniques which are commonly used in the mathematical and physical applications of tensor calculus. The chapter is made of three sections dedicated to tensors in mathematics, geometry and science. The mathematics part (see § 7.1) comes from tensor applications in linear algebra and vector calculus, while most of the materials in the science part (see § 7.3) come from applications related to fluid and continuum mechanics. In the geometry part (see § 7.2) a few prominent tensors and tensor identities of wide applications in differential geometry and related scientific fields are examined. We should also refer the reader to the previous chapter (see § 6) as an example of tensor applications in mathematics and science since the materials in that chapter are partly based on employing tensor notation and techniques. We note that all the aforementioned disciplines, where the materials about tensors come from, are intimately linked to tensor calculus since large parts of this subject were developed within those disciplines. Also, the materials presented in this chapter about tensors and tensor techniques are used as vital building blocks and tools in a number of important mathematical and physical theories. However, we would like to insist that although these materials provide a very useful glimpse, they are just partially representative examples of tensor applications. Our objective is to have more familiarity with some prominent tensors and tensor techniques and hence they are not meant to provide a comprehensive view. We should also indicate that some tensors, especially those in the science section, are defined differently in different disciplines and hence the given definitions and descriptions may not be thorough or universal. 7.1 Tensors in Mathematics In this section, we provide a sample of common definitions related to basic concepts and operations in matrix and vector algebra using the language of tensor calculus. Common identities in vector calculus as well as the integral theorems of Gauss and Stokes and some important scalar invariants are also examined in this section from this perspective. Finally, we provide a rather extensive set of examples about the use of tensor language and techniques in proving mathematical identities where these identities are gathered from vector calculus. For simplicity, clarity and wide applicability we employ a Cartesian approach in the tensor formulations of this section. 155 7.1.1 Common Definitions in Tensor Notation 7.1.1 156 Common Definitions in Tensor Notation The trace of a matrix A representing a rank-2 tensor in an nD space is given by: tr (A) = Aii (i = 1, . . . , n) (486) For a 3 × 3 matrix representing a rank-2 tensor in a 3D space, the determinant is given by: A11 A12 A13 det (A) = A21 A22 A23 = ǫijk A1i A2j A3k = ǫijk Ai1 Aj2 Ak3 (487) A31 A32 A33 where the last two equalities represent the expansion of the determinant by row and by column. Alternatively, the determinant of a 3 × 3 matrix can be given by: det (A) = 1 ǫijk ǫlmn Ail Ajm Akn 3! (488) More generally, for an n × n matrix representing a rank-2 tensor in an nD space, the determinant is given by: det (A) = ǫi1 ···in A1i1 . . . Anin = ǫi1 ···in Ai1 1 . . . Ain n 1 = ǫi ···i ǫj ···j Ai j . . . Ain jn n! 1 n 1 n 1 1 (489) The inverse of a matrix A representing a rank-2 tensor is given by:  A−1  ij = 1 ǫipq ǫjmn Amp Anq 2 det (A) (490) The multiplication of a matrix A by a vector b, as defined in linear algebra, is given by: [Ab]i = Aij bj (491) It should be remarked that we are using matrix notation in the writing of Ab. According to the symbolic notation of tensors, the multiplication operation should be denoted by a dot between the symbols of the tensor and the vector, i.e. A·b, since matrix multiplication in matrix algebra is equivalent to an inner product operation in tensor algebra. Similarly, the multiplication of two compatible matrices A and B, as defined in linear algebra, is given by: [AB]ik = Aij Bjk (492) Again, we are using here matrix notation in the writing of AB; otherwise a dot should be inserted between the symbols of the two matrices. The dot product of two vectors of the same dimension is given by: A · B = δij Ai Bj = Ai Bi (493) 7.1.2 Scalar Invariants of Tensors 157 Similarly, the cross product of two vectors in a 3D space is given by: [A × B]i = ǫijk Aj Bk (494) The scalar triple product of three vectors in a 3D space is given by: A · (B × C) = A1 A2 A3 B1 B2 B3 C1 C2 C3 = ǫijk Ai Bj Ck (495) while the vector triple product of three vectors in a 3D space is given by: [A × (B × C)]i = ǫijk ǫklm Aj Bl Cm (496) The expression of the other principal form of the vector triple product, i.e. (A × B) × C, can be obtained from the above form by changing the order of the factors in the external cross product and reversing the sign. Other operations, like relabeling the indices and exchanging some of the indices of the epsilons with a shift in sign, can then follow to obtain a more organized form. 7.1.2 Scalar Invariants of Tensors In the following, we list and write in tensor notation a number of invariants of low rank tensors which have special importance due to their widespread applications in vector and tensor calculus. All these invariants are scalars. The value of a scalar (rank-0 tensor), which consists of a magnitude and a sign, is invariant under coordinate transformations. An invariant of a vector (rank-1 tensor) under coordinate transformations is its magnitude, i.e. length.[29] The main three independent scalar invariants of a rank-2 tensor A are: I = tr (A) = Aii  II = tr A2 = Aij Aji  III = tr A3 = Aij Ajk Aki (497) (498) (499) I1 = I = Aii  1 1 2 I2 = I − II = (Aii Ajj − Aij Aji ) 2 2  1 1 3 I − 3I II + 2III = ǫijk ǫpqr Aip Ajq Akr I3 = det (A) = 3! 3! (500) Different forms of the three invariants of a rank-2 tensor A, which are also widely used, are the following (noting that some of these definitions may belong to 3D specifically): [29] (501) (502) The direction is also invariant but it is not a scalar! In fact the magnitude alone is invariant under coordinate transformations even for pseudo vectors because it is a true scalar. 7.1.3 Common Identities in Vector and Tensor Notation 158 where I, II and III are given by Eqs. 497-499. As the invariants I1 , I2 and I3 are defined in terms of the invariants I, II and III according to Eqs. 500-502, the invariants I, II and III can also be defined in terms of the invariants I1 , I2 and I3 , that is: I = I1 II = I12 − 2I2 III = I13 − 3I1 I2 + 3I3 (503) (504) (505) Since the determinant of a matrix representing a rank-2 tensor is invariant (see Eq. 502), then if the determinant vanishes in one coordinate system, it will vanish in all transformed coordinate systems, and if not it will not (also refer to § 3.1.4). Consequently, if a rank-2 tensor is invertible in a particular coordinate system, it will be invertible in all coordinate systems, and if not it will not. The following are ten common scalar invariants that are jointly formed between two rank2 tensors A and B: tr (A), tr (B), tr (A2 ), tr (B2 ), tr (A3 ), tr (B3 ), tr (A · B), tr (A2 · B), tr (A · B2 ) and tr (A2 · B2 ). As seen, all these are traces. 7.1.3 Common Identities in Vector and Tensor Notation In this subsection, we present some of the widely used identities of vector calculus using the traditional vector notation as well as its equivalent tensor notation. In the following equations, f and h are differentiable scalar fields; A, B, C and D are differentiable vector fields; and r = xi ei is the position vector. ∇·r = n ∂ i xi = n (vector notation) (tensor notation) (506) (507) where n is the space dimension. ∇×r = 0 ǫijk ∂j xk = 0 (508) (509) ∇ (a · r) = a ∂i (aj xj ) = ai (510) (511) ∇ · (∇f ) = ∇2 f ∂i (∂i f ) = ∂ii f (512) (513) ∇ · (∇ × A) = 0 ǫijk ∂i ∂j Ak = 0 (514) (515) where a is a constant vector. 7.1.3 Common Identities in Vector and Tensor Notation 159 ∇ × (∇f ) = 0 ǫijk ∂j ∂k f = 0 (516) (517) ∇ (f h) = f ∇h + h∇f ∂i (f h) = f ∂i h + h∂i f (518) (519) ∇ · (f A) = f ∇ · A + A · ∇f ∂i (f Ai ) = f ∂i Ai + Ai ∂i f (520) (521) ∇ × (f A) = f ∇ × A + ∇f × A ǫijk ∂j (f Ak ) = f ǫijk ∂j Ak + ǫijk (∂j f ) Ak (522) (523) A · (B × C) = C · (A × B) = B · (C × A) ǫijk Ai Bj Ck = ǫkij Ck Ai Bj = ǫjki Bj Ck Ai (524) (525) A × (B × C) = B (A · C) − C (A · B) ǫijk Aj ǫklm Bl Cm = Bi (Am Cm ) − Ci (Al Bl ) (526) (527) A × (∇ × B) = (∇B) · A − A · ∇B ǫijk ǫklm Aj ∂l Bm = (∂i Bm ) Am − Al (∂l Bi ) (528) (529) ∇ × (∇ × A) = ∇ (∇ · A) − ∇2 A ǫijk ǫklm ∂j ∂l Am = ∂i (∂m Am ) − ∂ll Ai (530) (531) ∇ (A · B) = A × (∇ × B) + B × (∇ × A) + (A · ∇) B + (B · ∇) A (532) ∂i (Am Bm ) = ǫijk Aj (ǫklm ∂l Bm ) + ǫijk Bj (ǫklm ∂l Am ) + (Al ∂l ) Bi + (Bl ∂l ) Ai (533) ∇ · (A × B) = B · (∇ × A) − A · (∇ × B) ∂i (ǫijk Aj Bk ) = Bk (ǫkij ∂i Aj ) − Aj (ǫjik ∂i Bk ) (534) (535) ∇ × (A × B) = (B · ∇) A + (∇ · B) A − (∇ · A) B − (A · ∇) B ǫijk ǫklm ∂j (Al Bm ) = (Bm ∂m ) Ai + (∂m Bm ) Ai − (∂j Aj ) Bi − (Aj ∂j ) Bi (536) (537) A·C A·D B·C B·D = (Al Cl ) (Bm Dm ) − (Am Dm ) (Bl Cl ) (A × B) · (C × D) = ǫijk Aj Bk ǫilm Cl Dm (538) (539) 7.1.4 Integral Theorems in Tensor Notation 160 (A × B) × (C × D) = [D · (A × B)] C − [C · (A × B)] D ǫijk ǫjmn Am Bn ǫkpq Cp Dq = (ǫqmn Dq Am Bn ) Ci − (ǫpmn Cp Am Bn ) Di (540) (541) In vector and tensor notations, the condition for a vector field A to be solenoidal is given by: (542) (543) ∇·A = 0 ∂i Ai = 0 In vector and tensor notations, the condition for a vector field A to be irrotational is given by: (544) (545) ∇×A = 0 ǫijk ∂j Ak = 0 7.1.4 Integral Theorems in Tensor Notation The divergence theorem for a differentiable vector field A in vector and tensor notations is given by: ¨ ˚ A · n dσ (546) ∇ · A dτ = S Ωˆ ˆ Ai ni dσ (547) ∂i Ai dτ = S Ω where Ω is a bounded region in an nD space enclosed by a generalized surface S, dτ and dσ are generalized volume and area differentials, n and ni are the unit vector normal to the surface and its ith component, and the index i ranges over 1, . . . , n. Similarly, the divergence theorem for a differentiable rank-2 tensor field A in tensor notation for the first index is given by: ˆ ˆ ∂i Ail dτ = Ail ni dσ (548) Ω S while the divergence theorem for differentiable tensor fields of higher rank A in tensor notation for the index k is given by: ˆ ˆ ∂k Aij...k...m dτ = Aij...k...m nk dσ (549) Ω S Stokes theorem for a differentiable vector field A in vector and tensor notations is given by: ¨ ˆ (∇ × A) · n dσ = A · dr (550) S C ˆ ˆ Ai dxi (551) ǫijk ∂j Ak ni dσ = S C 7.1.5 Examples of Using Tensor Techniques to Prove Identities 161 where C stands for the perimeter of the surface S, and dr is a differential of the position vector which is tangent to the perimeter while the other symbols are as defined above. Similarly, Stokes theorem for a differentiable rank-2 tensor field A in tensor notation for the first index is given by: ˆ ˆ Ail dxi (552) ǫijk ∂j Akl ni dσ = S C while Stokes theorem for differentiable tensor fields of higher rank A in tensor notation for the index k is given by: ˆ ˆ Alm...k...n dxk (553) ǫijk ∂j Alm...k...n ni dσ = S 7.1.5 C Examples of Using Tensor Techniques to Prove Identities In this subsection, we provide some examples for using tensor techniques to prove vector and tensor identities where comments are added next to each step to explain and justify. These examples, which are based on the identities given in § 7.1.3, demonstrate the elegance, efficiency and clarity of the methods and notation of tensor calculus. We note that in Cartesian coordinate systems, some tensor equations in general coordinates change their form and hence in the following we give the corresponding Cartesian form of the equations which are not given previously in this form since we use the Cartesian form in these proofs. The added comments inside the parentheses refer to the corresponding equations in general coordinates: ǫijk ǫlmk δij Aj ∂xi ∂xj ∂ i xi • ∇ · r = n: • ∇ × r = 0: = δil δjm − δim δjl = Ai (Eq. 167) (Eq. 172) (554) (555) = ∂j xi = xi,j = δij (Eq. 175) (556) = δii = n (Eq. 176) (557) ∇ · r = ∂ i xi = δii =n [∇ × r]i = ǫijk ∂j xk = ǫijk δkj = ǫijj =0 (Eq. 419) (Eq. 557) (Eq. 557) (Eq. (Eq. (Eq. (Eq. Since i is a free index, the identity is proved for all components. 423) 556) 555) 155) 7.1.5 Examples of Using Tensor Techniques to Prove Identities • ∇ (a · r) = a: [∇ (a · r)]i = ∂i (aj xj ) = aj ∂i xj + xj ∂i aj = aj ∂i xj = aj δji = ai = [a]i 162 (Eqs. 417 & 493) (product rule) (aj is constant) (Eq. 556) (Eq. 555) (definition of index) Since i is a free index, the identity is proved for all components. • ∇ · (∇f ) = ∇2 f : ∇ · (∇f ) = ∂i [∇f ]i (Eq. 419) = ∂i (∂i f ) (Eq. 417) = ∂i ∂i f (rules of differentiation) = ∂ii f (definition of 2nd derivative) 2 =∇ f (Eq. 425) • ∇ · (∇ × A) = 0: ∇ · (∇ × A) = ∂i [∇ × A]i = ∂i (ǫijk ∂j Ak ) = ǫijk ∂i ∂j Ak = ǫijk ∂j ∂i Ak = −ǫjik ∂j ∂i Ak = −ǫijk ∂i ∂j Ak =0 (Eq. 419) (Eq. 423) (ǫ is constant) (continuity condition) (Eq. 180) (relabeling dummy indices i and j) (since ǫijk ∂i ∂j Ak = −ǫijk ∂i ∂j Ak ) This can also be concluded from line three by arguing that: since by the continuity condition ∂i and ∂j can change their order with no change in the value of the term while a corresponding change of the order of i and j in ǫijk results in a sign change, we see that each term in the sum has its own negative and hence the terms add up to zero. • ∇ × (∇f ) = 0: [∇ × (∇f )]i = ǫijk ∂j [∇f ]k (Eq. 423) = ǫijk ∂j (∂k f ) (Eq. 417) = ǫijk ∂j ∂k f (rules of differentiation) = ǫijk ∂k ∂j f (continuity condition) = −ǫikj ∂k ∂j f (Eq. 180) = −ǫijk ∂j ∂k f (relabeling dummy indices j and k) =0 (since ǫijk ∂j ∂k f = −ǫijk ∂j ∂k f ) This can also be concluded from line three by a similar argument to the one given in the previous point. Because [∇ × (∇f )]i is an arbitrary component, then each component is zero. 7.1.5 Examples of Using Tensor Techniques to Prove Identities • ∇ (f h) = f ∇h + h∇f : [∇ (f h)]i = ∂i (f h) = f ∂i h + h∂i f = [f ∇h]i + [h∇f ]i = [f ∇h + h∇f ]i 163 (Eq. 417) (product rule) (Eq. 417) (Eq. 24) Because i is a free index, the identity is proved for all components. • ∇ · (f A) = f ∇ · A + A · ∇f : ∇ · (f A) = ∂i [f A]i (Eq. 419) = ∂i (f Ai ) (definition of index) = f ∂i Ai + Ai ∂i f (product rule) = f ∇ · A + A · ∇f (Eqs. 419 & 413) • ∇ × (f A) = f ∇ × A + ∇f × A: [∇ × (f A)]i = ǫijk ∂j [f A]k = ǫijk ∂j (f Ak ) = f ǫijk ∂j Ak + ǫijk (∂j f ) Ak = f ǫijk ∂j Ak + ǫijk [∇f ]j Ak = [f ∇ × A]i + [∇f × A]i = [f ∇ × A + ∇f × A]i (Eq. 423) (definition of index) (product rule & commutativity) (Eq. 417) (Eqs. 423 & 494) (Eq. 24) Because i is a free index, the identity is proved for all components. • A · (B × C) = C · (A × B) = B · (C × A): A · (B × C) = ǫijk Ai Bj Ck (Eq. 495) = ǫkij Ai Bj Ck (Eq. 180) = ǫkij Ck Ai Bj (commutativity) = C · (A × B) (Eq. 495) = ǫjki Ai Bj Ck (Eq. 180) = ǫjki Bj Ck Ai (commutativity) = B · (C × A) (Eq. 495) The negative permutations of this identity can be similarly obtained and proved by changing the order of the vectors in the cross products which results in a sign change. • A × (B × C) = B (A · C) − C (A · B): [A × (B × C)]i = ǫijk Aj [B × C]k (Eq. 494) = ǫijk Aj ǫklm Bl Cm (Eq. 494) = ǫijk ǫklm Aj Bl Cm (commutativity) = ǫijk ǫlmk Aj Bl Cm (Eq. 180) = (δil δjm − δim δjl ) Aj Bl Cm (Eq. 554) 7.1.5 Examples of Using Tensor Techniques to Prove Identities = δil δjm Aj Bl Cm − δim δjl Aj Bl Cm = (δil Bl ) (δjm Aj Cm ) − (δim Cm ) (δjl Aj Bl ) = Bi (Am Cm ) − Ci (Al Bl ) = Bi (A · C) − Ci (A · B) = [B (A · C)]i − [C (A · B)]i = [B (A · C) − C (A · B)]i 164 (distributivity) (commutativity & grouping) (Eq. 555) (Eq. 493) (definition of index) (Eq. 24) Because i is a free index, the identity is proved for all components. The other variant of this identity, i.e. (A × B) × C, can be obtained and proved similarly by changing the order of the factors in the external cross product with adding a minus sign. • A × (∇ × B) = (∇B) · A − A · ∇B: [A × (∇ × B)]i = ǫijk Aj [∇ × B]k (Eq. 494) = ǫijk Aj ǫklm ∂l Bm (Eq. 423) = ǫijk ǫklm Aj ∂l Bm (commutativity) = ǫijk ǫlmk Aj ∂l Bm (Eq. 180) = (δil δjm − δim δjl ) Aj ∂l Bm (Eq. 554) = δil δjm Aj ∂l Bm − δim δjl Aj ∂l Bm (distributivity) = Am ∂i Bm − Al ∂l Bi (Eq. 555) = (∂i Bm ) Am − Al (∂l Bi ) (commutativity & grouping) = [(∇B) · A]i − [A · ∇B]i (Eq. 418 & § 3.2.5) = [(∇B) · A − A · ∇B]i (Eq. 24) Because i is a free index, the identity is proved for • ∇ × (∇ × A) = ∇ (∇ · A) − ∇2 A: [∇ × (∇ × A)]i = ǫijk ∂j [∇ × A]k = ǫijk ∂j (ǫklm ∂l Am ) = ǫijk ǫklm ∂j (∂l Am ) = ǫijk ǫlmk ∂j ∂l Am = (δil δjm − δim δjl ) ∂j ∂l Am = δil δjm ∂j ∂l Am − δim δjl ∂j ∂l Am = ∂m ∂i Am − ∂l ∂l Ai = ∂i (∂m Am ) − ∂ll Ai   = [∇ (∇ · A)]i − ∇2 A i   = ∇ (∇ · A) − ∇2 A i all components. (Eq. 423) (Eq. 423) (ǫ is constant) (Eq. 180 & definition of derivative) (Eq. 554) (distributivity) (Eq. 555) (∂ shift, grouping & Eq. 2) (Eqs. 419, 417 & 426) (Eq. 24) Because i is a free index, the identity is proved for all components. This identity can also be considered as an instance of the identity before the last one, observing that in the second term on the right hand side the Laplacian should precede the vector, and hence no independent proof is required. 7.1.5 Examples of Using Tensor Techniques to Prove Identities • ∇ (A · B) = A × (∇ × B) + B × (∇ × A) + (A · ∇) B + (B · ∇) A: We start from the right hand side and end with the left hand side: [A × (∇ × B) + B × (∇ × A) + (A · ∇) B + (B · ∇) A]i = [A × (∇ × B)]i + [B × (∇ × A)]i + [(A · ∇) B]i + [(B · ∇) A]i = (Eq. 24) ǫijk Aj [∇ × B]k + ǫijk Bj [∇ × A]k + (Al ∂l ) Bi + (Bl ∂l ) Ai = (Eqs. 494, 419 & indexing) ǫijk Aj (ǫklm ∂l Bm ) + ǫijk Bj (ǫklm ∂l Am ) + (Al ∂l ) Bi + (Bl ∂l ) Ai = (Eq. 423) ǫijk ǫklm Aj ∂l Bm + ǫijk ǫklm Bj ∂l Am + (Al ∂l ) Bi + (Bl ∂l ) Ai = (commutativity) ǫijk ǫlmk Aj ∂l Bm + ǫijk ǫlmk Bj ∂l Am + (Al ∂l ) Bi + (Bl ∂l ) Ai = (Eq. 180) (δil δjm − δim δjl ) Aj ∂l Bm + (δil δjm − δim δjl ) Bj ∂l Am + (Al ∂l ) Bi + (Bl ∂l ) Ai = (Eq. 554) (δil δjm Aj ∂l Bm − δim δjl Aj ∂l Bm ) + (δil δjm Bj ∂l Am − δim δjl Bj ∂l Am ) + (Al ∂l ) Bi + (Bl ∂l ) Ai = (distributivity) δil δjm Aj ∂l Bm − Al ∂l Bi + δil δjm Bj ∂l Am − Bl ∂l Ai + (Al ∂l ) Bi + (Bl ∂l ) Ai = (Eq. 555) δil δjm Aj ∂l Bm − (Al ∂l ) Bi + δil δjm Bj ∂l Am − (Bl ∂l ) Ai + (Al ∂l ) Bi + (Bl ∂l ) Ai = (grouping) δil δjm Aj ∂l Bm + δil δjm Bj ∂l Am = (cancellation) Am ∂i Bm + Bm ∂i Am = (Eq. 555) ∂i (Am Bm ) = (product rule) [∇ (A · B)]i (Eqs. 417 & 419) Because i is a free index, the identity is proved for all components. • ∇ · (A × B) = B · (∇ × A) − A · (∇ × B): ∇ · (A × B) = ∂i [A × B]i (Eq. 419) = ∂i (ǫijk Aj Bk ) (Eq. 494) = ǫijk ∂i (Aj Bk ) (ǫ is constant) = ǫijk (Bk ∂i Aj + Aj ∂i Bk ) (product rule) 165 7.1.5 Examples of Using Tensor Techniques to Prove Identities = ǫijk Bk ∂i Aj + ǫijk Aj ∂i Bk = ǫkij Bk ∂i Aj − ǫjik Aj ∂i Bk = Bk (ǫkij ∂i Aj ) − Aj (ǫjik ∂i Bk ) = Bk [∇ × A]k − Aj [∇ × B]j = B · (∇ × A) − A · (∇ × B) 166 (distributivity) (Eq. 180) (commutativity & grouping) (Eq. 423) (Eq. 493) • ∇ × (A × B) = (B · ∇) A + (∇ · B) A − (∇ · A) B − (A · ∇) B: [∇ × (A × B)]i = ǫijk ∂j [A × B]k = ǫijk ∂j (ǫklm Al Bm ) = ǫijk ǫklm ∂j (Al Bm ) = ǫijk ǫklm (Bm ∂j Al + Al ∂j Bm ) = ǫijk ǫlmk (Bm ∂j Al + Al ∂j Bm ) = (δil δjm − δim δjl ) (Bm ∂j Al + Al ∂j Bm ) = δil δjm Bm ∂j Al + δil δjm Al ∂j Bm − δim δjl Bm ∂j Al − δim δjl Al ∂j Bm = Bm ∂m Ai + Ai ∂m Bm − Bi ∂j Aj − Aj ∂j Bi = (Bm ∂m ) Ai + (∂m Bm ) Ai − (∂j Aj ) Bi − (Aj ∂j ) Bi = [(B · ∇) A]i + [(∇ · B) A]i − [(∇ · A) B]i − [(A · ∇) B]i = [(B · ∇) A + (∇ · B) A − (∇ · A) B − (A · ∇) B]i (Eq. 423) (Eq. 494) (ǫ is constant) (product rule) (Eq. 180) (Eq. 554) (distributivity) (Eq. 555) (grouping) (Eqs. 413 & 419) (Eq. 24) Because i is a free index, the identity is proved for all components. A·C A·D : B·C B·D (A × B) · (C × D) = [A × B]i [C × D]i = ǫijk Aj Bk ǫilm Cl Dm = ǫijk ǫilm Aj Bk Cl Dm = (δjl δkm − δjm δkl ) Aj Bk Cl Dm = δjl δkm Aj Bk Cl Dm − δjm δkl Aj Bk Cl Dm = (δjl Aj Cl ) (δkm Bk Dm ) − (δjm Aj Dm ) (δkl Bk Cl ) = (Al Cl ) (Bm Dm ) − (Am Dm ) (Bl Cl ) = (A · C) (B · D) − (A · D) (B · C) • (A × B) · (C × D) = = A·C A·D B·C B·D • (A × B) × (C × D) = [D · (A × B)] C − [C · (A × B)] D: [(A × B) × (C × D)]i = ǫijk [A × B]j [C × D]k = ǫijk ǫjmn Am Bn ǫkpq Cp Dq (Eq. 493) (Eq. 494) (commutativity) (Eqs. 180 & 554) (distributivity) (commutativity) (Eq. 555) (Eq. 493) (definition) (Eq. 494) (Eq. 494) 7.2 Tensors in Geometry 167 = ǫijk ǫkpq ǫjmn Am Bn Cp Dq = ǫijk ǫpqk ǫjmn Am Bn Cp Dq = (δip δjq − δiq δjp ) ǫjmn Am Bn Cp Dq = (δip δjq ǫjmn − δiq δjp ǫjmn ) Am Bn Cp Dq = (δip ǫqmn − δiq ǫpmn ) Am Bn Cp Dq = δip ǫqmn Am Bn Cp Dq − δiq ǫpmn Am Bn Cp Dq = ǫqmn Am Bn Ci Dq − ǫpmn Am Bn Cp Di = ǫqmn Dq Am Bn Ci − ǫpmn Cp Am Bn Di = (ǫqmn Dq Am Bn ) Ci − (ǫpmn Cp Am Bn ) Di = [D · (A × B)] Ci − [C · (A × B)] Di = [[D · (A × B)] C]i − [[C · (A × B)] D]i = [[D · (A × B)] C − [C · (A × B)] D]i (commutativity) (Eq. 180) (Eq. 554) (distributivity) (Eq. 555) (distributivity) (Eq. 555) (commutativity) (grouping) (Eq. 495) (index definition) (Eq. 24) Because i is a free index, the identity is proved for all components. 7.2 Tensors in Geometry In this section, we give some examples of tensor calculus applications in geometry. These examples are related mainly to the investigation of the properties of spaces in general and space curvature in particular and hence they play important roles in non-Euclidean geometries and related applications in the physical sciences such as geometry-based gravitational theories. 7.2.1 Riemann-Christoffel Curvature Tensor This absolute rank-4 tensor, which is also called Riemann curvature tensor and RiemannChristoffel tensor, is a property of the space. It characterizes important properties of spaces and hence it plays an important role in geometry in general and in non-Euclidean geometries in particular with many applications in geometry-based physical theories. The tensor is used, for instance, to test for the space flatness (see § 2.1). As indicated before, the Riemann-Christoffel curvature tensor vanishes identically iff the space is globally flat and hence the Riemann-Christoffel curvature tensor is zero in Euclidean spaces. The Riemann-Christoffel curvature tensor depends only on the metric which, in general coordinate systems, is a function of position and hence the Riemann-Christoffel curvature tensor follows this dependency on position. Yes, for affine coordinate systems the metric tensor is constant and hence the Riemann-Christoffel curvature tensor vanishes identically. There are two kinds of Riemann-Christoffel curvature tensor: first and second, where the first kind is a type (0, 4) tensor while the second kind is a type (1, 3) tensor. The RiemannChristoffel curvature tensor of the first kind, which may also be called the covariant (or totally covariant) Riemann-Christoffel curvature tensor, is defined by: Rijkl = ∂k [jl, i] − ∂l [jk, i] + [il, r] Γrjk − [ik, r] Γrjl (558) 7.2.1 Riemann-Christoffel Curvature Tensor 168 1 (∂k ∂j gli + ∂l ∂i gjk − ∂k ∂i gjl − ∂l ∂j gki ) + [il, r] Γrjk − [ik, r] Γrjl 2 1 (∂k ∂j gli + ∂l ∂i gjk − ∂k ∂i gjl − ∂l ∂j gki ) + g rs ([il, r] [jk, s] − [ik, r] [jl, s]) = 2 = where the second step is based on Eq. 355 while the third step is based on Eq. 308. We note that the first line of Eq. 558 can be cast in the following mnemonic determinantal form: Γrjk Γrjl ∂k ∂l Rijkl = + (559) [jk, i] [jl, i] [ik, r] [il, r] Similarly, the Riemann-Christoffel curvature tensor of the second kind, which may also be called the mixed Riemann-Christoffel curvature tensor, is given by: Rijkl = ∂k Γijl − ∂l Γijk + Γrjl Γirk − Γrjk Γirl (560) We note again that Eq. 560 can be put into the following mnemonic determinantal form: Rijkl = ∂k ∂l Γijk Γijl + Γrjl Γrjk Γirl Γirk (561) On lowering the contravariant index of the Riemann-Christoffel curvature tensor of the second kind, the Riemann-Christoffel curvature tensor of the first kind is obtained, that is: Rijkl = gia Rajkl (562) Similarly, the Riemann-Christoffel curvature tensor of the second kind can be obtained by raising the first index of the Riemann-Christoffel curvature tensor of the first kind, that is: Ri jkl = g ia Rajkl (563) One of the main applications of the Riemann-Christoffel curvature tensor in tensor calculus is demonstrated in its role in tensor differentiation. As seen in § 5.2, the covariant differential operators in mixed derivatives are not commutative and hence for a covariant vector A we have: Aj;kl − Aj;lk = Rijkl Ai (564) From Eq. 564, it is obvious that the mixed second order covariant derivatives are equal iff the Riemann-Christoffel curvature tensor of the second kind vanishes identically which means that the space is flat and hence the covariant derivatives are ordinary partial derivatives. Similarly, for the mixed second order covariant derivative of a contravariant vector A we have: (565) Aj;kl − Aj;lk = Rjilk Ai which is similar to Eq. 564 for a covariant vector A. In brief, the covariant differential operators become commutative when the metric makes the Riemann-Christoffel curvature tensor of either kind vanish. So, since the Riemann-Christoffel curvature tensor is zero in 7.2.1 Riemann-Christoffel Curvature Tensor 169 Euclidean spaces, the mixed second order covariant derivatives, which become ordinary partial derivatives, are equal when the C 2 continuity condition is satisfied. The Riemann-Christoffel curvature tensor of the first kind satisfies the following symmetric and skew-symmetric relations in its four indices: Rijkl = Rklij = −Rjikl = −Rijlk (block symmetry) (anti-symmetry in the first two indices) (anti-symmetry in the last two indices) (566) The skew-symmetric property of the covariant Riemann-Christoffel curvature tensor with respect to the last two indices also applies to the mixed Riemann-Christoffel curvature tensor, that is: Rijkl = −Rijlk (567) As a consequence of the two anti-symmetric properties of the covariant Riemann-Christoffel curvature tensor, the entries of the Riemann-Christoffel curvature tensor with identical values of the first two indices (e.g. Riijk ) or/and the last two indices (e.g. Rijkk ) are zero (refer to § 3.1.5). Also, as a consequence of these anti-symmetric properties, all the entries of the tensor with identical values of more than two indices (e.g. Riiji ) are zero. We remark that all the above symmetric and anti-symmetric properties of the RiemannChristoffel curvature tensor of the first and second kinds can be proved by using the above definitions of this tensor. For example, the first anti-symmetric property can be verified as follows: 1 (∂k ∂i glj + ∂l ∂j gik − ∂k ∂j gil − ∂l ∂i gkj ) + g rs ([jl, r] [ik, s] − [jk, r] [il, s]) (568) Rjikl = 2  1 rs = − (∂k ∂j gil + ∂l ∂i gkj − ∂k ∂i glj − ∂l ∂j gik ) + g ([jk, r] [il, s] − [jl, r] [ik, s]) 2   1 sr (∂k ∂j gli + ∂l ∂i gjk − ∂k ∂i gjl − ∂l ∂j gki ) + g ([il, s] [jk, r] − [ik, s] [jl, r]) = − 2   1 rs (∂k ∂j gli + ∂l ∂i gjk − ∂k ∂i gjl − ∂l ∂j gki ) + g ([il, r] [jk, s] − [ik, r] [jl, s]) = − 2 = −Rijkl where in the first line we use the third line of Eq. 558 with exchanging i and j, in the second line we take a factor of −1, in the third line we use the symmetry of the metric tensor in its two indices, in the fourth line we relabel two dummy indices, and in the last line we use the third line of Eq. 558 again. In an nD space, the Riemann-Christoffel curvature tensor has n4 components. As a consequence of the symmetric and anti-symmetric properties of the covariant RiemannChristoffel curvature tensor, there are three types of distinct non-vanishing entries: 1. Entries with only two distinct indices (type Rijij ) which count: N2 = n (n − 1) 2 (569) 7.2.1 Riemann-Christoffel Curvature Tensor 170 2. Entries with only three distinct indices (type Rijik ) which count: N3 = n (n − 1) (n − 2) 2 (570) 3. Entries with four distinct indices (type Rijkl ) which count: N4 = n (n − 1) (n − 2) (n − 3) 12 (571) The numerator in these three equations represents the k-permutations of n distinct objects which is given by: n! P (n, k) = (572) (n − k)! where k in these cases is equal to 2, 3, 4 respectively, while the denominator represents the number of non-independent ways of generating these entries due to the anti-symmetric and block symmetric properties. By adding the numbers of the three types of non-zero distinct entries, as given by Eqs. 569-571, we can see that the Riemann-Christoffel curvature tensor in an nD space has a total of: n2 (n2 − 1) NRI = N2 + N3 + N4 = (573) 12 independent components which do not vanish identically. For example, in a 2D Riemannian space the Riemann-Christoffel curvature tensor has 24 = 16 components; however there is only one independent component (with the principal suffix 1212) which is not identically zero represented by the following four interdependent components: R1212 = R2121 = −R1221 = −R2112 (574) Similarly, in a 3D Riemannian space the Riemann-Christoffel curvature tensor has 34 = 81 components but only six of these are distinct non-zero entries which are the ones with the following principal suffixes: 1212, 1313, 1213, 2123, 3132, 2323 (575) where the permutations of the indices in each one of these suffixes are subject to the symmetric and anti-symmetric properties of the four indices of the Riemann-Christoffel curvature tensor, as in the case of a 2D space in the above example, and hence these permutations do not produce independent entries. Following the pattern in the examples of 2D and 3D spaces, in a 4D Riemannian space the Riemann-Christoffel curvature tensor has 44 = 256 components but only 20 of these are independent non-zero entries, while in a 5D Riemannian space the Riemann-Christoffel curvature tensor has 54 = 625 components but only 50 are independent non-zero entries. On contracting the first covariant index with the contravariant index of the RiemannChristoffel curvature tensor of the second kind we obtain: Riikl = ∂k Γiil − ∂l Γiik + Γril Γirk − Γrik Γirl (j = i in Eq. 560) (576) 7.2.2 Bianchi Identities 171 = ∂k Γiil − ∂l Γiik + Γril Γirk − Γirk Γril ∂k Γiil (relabeling dummy i, r in last term) ∂l Γiik − = √ √ = ∂k [∂l (ln g)] − ∂l [∂k (ln g)] √ √ = ∂k ∂l (ln g) − ∂l ∂k (ln g) √ √ = ∂k ∂l (ln g) − ∂k ∂l (ln g) =0 (Eq. 324) (C 2 continuity condition) That is: Riikl = 0 (577) This is inline with the anti-symmetric property of the first two indices of the totally covariant Riemann-Christoffel curvature tensor. Finally, the Riemann-Christoffel curvature tensor also satisfies the following identity: (578) Rijkl;s + Riljk;s = Riksl;j + Rikjs;l As indicated earlier, a necessary and sufficient condition for a space to have a coordinate system such that all the components of the metric tensor are constants is that: (579) Rijkl = 0 This can be concluded from Eq. 558 plus Eqs. 307 and 308. 7.2.2 Bianchi Identities The Riemann-Christoffel curvature tensor of the first and second kind satisfies a number of identities called the Bianchi identities. The first Bianchi identity is given by: Rijkl + Riljk + Riklj = 0 (first kind) (580) Rijkl + Riljk + Riklj = 0 (second kind) (581) These two forms of the first Bianchi identity can be obtained from each other by the index raising and lowering operators. The first Bianchi identity, as stated above, is an instance of the fact that by fixing the position of one of the four indices and permuting the other three indices cyclically, the algebraic sum of these three permuting forms is zero, that is: Rijkl + Riljk + Riklj Rijkl + Rljik + Rkjli Rijkl + Rlikj + Rjlki Rijkl + Rkijl + Rjkil =0 =0 =0 =0 (i fixed) (j fixed) (k fixed) (l fixed) (582) (583) (584) (585) All the above identities can be easily proved by using the definition of the RiemannChristoffel curvature tensor noting that the Christoffel symbols of both kinds are symmetric in their paired indices as given by Eqs. 310 and 311. For example, Eq. 580 can be verified by substituting from the first line of Eq. 558 into Eq. 580, that is: Rijkl + Riljk + Riklj = ∂k [jl, i] − ∂l [jk, i] + [il, r] Γrjk − [ik, r] Γrjl + (586) 7.2.3 Ricci Curvature Tensor and Scalar 172 ∂j [lk, i] − ∂k [lj, i] + [ik, r] Γrlj − [ij, r] Γrlk + ∂l [kj, i] − ∂j [kl, i] + [ij, r] Γrkl − [il, r] Γrkj = 0 Another one of the Bianchi identities is: Rijkl;m + Rijlm;k + Rijmk;l = 0 (first kind) (587) Rijkl;m + Rijlm;k + Rijmk;l = 0 (second kind) (588) Again, these two forms can be obtained from each other by the raising and lowering operators. We note that the pattern of the second Bianchi identity in its both kinds is simple, that is the first two indices are fixed while the last three indices are cyclically permuted in the three terms. It is noteworthy that the Bianchi identities are valid regardless of the space metric. 7.2.3 Ricci Curvature Tensor and Scalar The Ricci curvature tensor, which is a rank-2 absolute symmetric tensor, is a byproduct of the Riemann-Christoffel curvature tensor and hence it plays a similar role in characterizing the space and describing its curvature. There are two kinds of Ricci curvature tensor: first and second, where the first kind is a type (0, 2) tensor while the second kind is a type (1, 1) tensor. The first kind of this tensor is obtained by contracting the contravariant index with the last covariant index of the Riemann-Christoffel curvature tensor of the second kind, that is: (589) Rij = Raija = ∂j Γaia − ∂a Γaij + Γabj Γbia − Γaba Γbij where Eq. 560 is used in this formulation. The Ricci curvature tensor, as given by the last equation, can be written in the following mnemonic determinantal form: Γabj Γaba ∂j ∂a (590) + Γaij Γaia Γbij Γbia √  Because of Eq. 324 (i.e. Γjij = ∂i ln g ), the Ricci tensor can also be written in the following forms as well as several other forms: √ √ (591) Rij = ∂j ∂i (ln g) − ∂a Γaij + Γabj Γbia − Γbij ∂b (ln g) 1 √ √ a = ∂j ∂i (ln g) + Γabj Γbia − √ ∂a gΓij g Rij = where g is the determinant of the covariant metric tensor. As stated above, the Ricci tensor of the first kind is symmetric, that is: Rij = Rji (592) This can be easily verified by exchanging the i and j indices in the last line of Eq. 591 taking account of the C 2 continuity condition and the fact that the Christoffel symbols are symmetric in their paired indices (Eq. 311). 7.3 Tensors in Science 173 On raising the first index of the Ricci tensor of the first kind, the Ricci tensor of the second kind is obtained, that is: Rij = g ik Rkj (593) This process can be reversed and hence the Ricci tensor of the first kind can be obtained from the Ricci tensor of the second kind using the index lowering operator. In an nD space, the Ricci tensor has n2 entries. However, because of its symmetry the number of its distinct entries is reduced to: n (n + 1) (594) NRD = 2 The Ricci scalar R, which is also called the curvature scalar and the curvature invariant, is the result of contracting the indices of the Ricci tensor of the second kind, that is: R = Rii (595) Since the Ricci scalar is obtained by raising a covariant index of the Ricci tensor of the first kind using the raising operator followed by contracting the two indices, it may be written by some in the following form:   1 √ a √ a b ij ij (596) gΓij R = g Rij = g ∂j ∂i (ln g) + Γbj Γia − √ ∂a g where the expression inside the square brackets is obtained from the last line of Eq. 591. Similar expressions can be obtained from the other lines of that equation. 7.3 Tensors in Science In this section, we give a few examples of tensor calculus applications in science. These examples come mainly from the disciplines of continuum mechanics and fluid dynamics. For simplicity, clarity and widespread use we employ a Cartesian approach in the following formulations. 7.3.1 Infinitesimal Strain Tensor This is a rank-2 tensor which describes the state of strain in a continuum medium and hence it is used in continuum and fluid mechanics. The infinitesimal strain tensor γ is defined by: ∇d + (∇d)T γ= (597) 2 where d is the displacement vector and the superscript T represents matrix transposition. We note that some authors do not include the factor 21 in the definition of γ. The displacement vector d represents the change in distance and direction which an infinitesimal element of the medium experiences as a consequence of the applied stress. In Cartesian coordinates, the last equation is given in tensor notation by the following form: γij = ∂i dj + ∂j di 2 (598) 7.3.2 Stress Tensor 7.3.2 174 Stress Tensor The stress tensor, which is also called Cauchy stress tensor, is a rank-2 tensor used for transforming a normal vector to a surface to a traction vector acting on that surface, that is: T=σ·n (599) where T is the traction vector, σ is the stress tensor and n is the normal vector. This is usually expressed in tensor notation using Cartesian coordinates by the following form: (600) Ti = σij nj We should remark that the stress tensor is symmetric in many applications (e.g. in the flow of Newtonian fluids) but not all, as it can be asymmetric in some cases. We also remark that we chose to define the stress tensor within the context of Cauchy stress law which is more relevant to the continuum mechanics, although it can be defined differently in other disciplines and in a more general form. The diagonal components of the stress tensor represent normal stresses while the offdiagonal components represent shear stresses. Assuming that the stress tensor is symindependent components, instead of n2 , where n is the space metric, it possesses n(n+1) 2 dimension. Hence in a 3D space (which is the natural space for this tensor in the common physical applications) it has six independent components. In fluid dynamics, the stress tensor (which may also be labeled as the total stress tensor) is decomposed into two main parts: a viscous part and a pressure part. The viscous part may then be split into a normal stress and a shear stress while the pressure part may be split into a hydrostatic pressure and an extra pressure of hydrodynamic nature. 7.3.3 Displacement Gradient Tensors These are rank-2 tensors denoted by E and ∆. They are defined in Cartesian coordinates using tensor notation as: ′ ∂xi ∆ij = ∂xj ∂xi Eij = ′ ∂xj (601) ′ where the indexed x and x represent the Cartesian coordinates of an observed continuum particle at the present and past times respectively. These tensors may also be called the deformation gradient tensors. The first displacement gradient tensor E quantifies the displacement of a particle at the present time relative to its position at the past time, while the second displacement gradient tensor ∆ quantifies its displacement at the past time relative to its position at the present time. From their definitions, it is obvious that E and ∆ are inverses of each other and hence: Eik ∆kj = δij (602) 7.3.4 Finger Strain Tensor 7.3.4 175 Finger Strain Tensor This rank-2 tensor, which may also be called the left Cauchy-Green deformation tensor, is used in the fields of fluid and continuum mechanics to describe the strain in a continuum object, e.g. fluid, in a series of time frames. It is defined as: B = E · ET (603) where E is the first displacement gradient tensor as defined in § 7.3.3 and the superscript T represents matrix transposition. The last equation can be expressed in tensor notation, using Cartesian coordinates, as follows: Bij = ∂xi ∂xj ′ ′ ∂xk ∂xk (604) ′ where the indexed x and x represent the Cartesian coordinates of an element of the continuum at the present and past times respectively. 7.3.5 Cauchy Strain Tensor This tensor, which may also be called the right Cauchy-Green deformation tensor, is the inverse of the Finger strain tensor (see § 7.3.4) and hence it is denoted by B−1 . Consequently, it is defined as: B−1 = ∆T · ∆ (605) where ∆ is the second displacement gradient tensor as defined in § 7.3.3. The last equation can be expressed in tensor notation, using Cartesian coordinates, as follows: ′ Bij−1 ′ ∂xk ∂xk = ∂xi ∂xj (606) The Finger and Cauchy strain tensors may be called “finite strain tensors” as opposite to infinitesimal strain tensor (see § 7.3.1). They are symmetric positive definite tensors; moreover they become the unity tensor when the change in the state of the object from the past to the present times consists of rotation and translation with no deformation. 7.3.6 Velocity Gradient Tensor This is a rank-2 tensor which is often used in fluid dynamics and rheology. As its name suggests, it is the gradient of the velocity vector v and hence it is given in Cartesian coordinates by: [∇v]ij = ∂i vj (607) The velocity gradient tensor in other coordinate systems can be obtained from the expression of the gradient of vectors in these systems, as given, for instance, in § 6.4 and § 6.5 for cylindrical and spherical coordinates. The term “velocity gradient tensor” may also be used for the transpose of this tensor, i.e. (∇v)T . The velocity gradient tensor is usually 7.3.7 Rate of Strain Tensor 176 decomposed into a symmetric part which is the rate of strain tensor S (see § 7.3.7), and an anti-symmetric part which is the vorticity tensor S̄ (see § 7.3.8), that is: ∇v = S + S̄ (608) As seen earlier (refer to § 3.1.5), any rank-2 tensor can be decomposed into a symmetric part and an anti-symmetric part. 7.3.7 Rate of Strain Tensor This rank-2 tensor, which is also called the rate of deformation tensor, is the symmetric part of the velocity gradient tensor and hence it is given by: ∇v + (∇v)T 2 which, in tensor notation with Cartesian coordinates, becomes (see Eq. 99): S= (609) ∂ i vj + ∂ j vi (610) 2 We note that some authors do not include the factor 21 in the definition of S and S̄ and hence this factor is moved to the definition of ∇v (Eq. 608). Also, the tensors S and S̄ are commonly denoted by γ̇ and ω respectively. The rate of strain tensor is a quantitative measure of the local rate at which neighboring material elements of a deforming continuum move with respect to each other. As a rank-2 symmetric tensor, it has n(n+1) independent components which are six in a 3D space. The 2 rate of strain tensor is related to the infinitesimal strain tensor (refer to § 7.3.1) by: Sij = ∂γ (611) ∂t where t is time. Hence, the rate of strain tensor is normally denoted by γ̇ where the dot represents the temporal rate of change, as indicated above. S= 7.3.8 Vorticity Tensor This rank-2 tensor is the anti-symmetric part of the velocity gradient tensor and hence it is given by: ∇v − (∇v)T (612) S̄ = 2 which, in tensor notation with Cartesian coordinates, becomes (see Eq. 100): ∂ i vj − ∂ j vi (613) 2 The vorticity tensor quantifies the local rate of rotation of a deforming continuum medium. As a rank-2 anti-symmetric tensor, it has n(n−1) independent non-zero components which 2 are three in a 3D space. These three components, added to the aforementioned six components of the rate of strain tensor (see § 7.3.7), give nine independent components which is the total number of components of their parent tensor ∇v. S̄ij = 7.4 Exercises and Revision 7.4 177 Exercises and Revision 7.1 Summarize the reasons for the popularity of tensor calculus techniques in mathematical, scientific and engineering applications. 7.2 State, in tensor language, the definition of the following mathematical concepts assuming Cartesian coordinates of a 3D space: trace of matrix, determinant of matrix, inverse of matrix, multiplication of two compatible square matrices, dot product of two vectors, cross product of two vectors, scalar triple product of three vectors and vector triple product of three vectors. 7.3 From the tensor definition of A×(B × C), obtain the tensor definition of (A × B)×C. 7.4 We have the following tensors in orthonormal Cartesian coordinates of a 3D space: A = (22, 3π, 6.3) B = (3e, 1.8, 4.9) C = (47, 5e, 3.5)     3 e2 π 3 E= D= π3 7 4 e 7.5 7.6 7.7 7.8 Use the tensor expressions for the relevant mathematical concepts with systematic substitution of the indices values to find the following: tr (D), det (E), D−1 , E · D, A · C, C × B, C · (A × B) and B × (C × A). State the matrix and tensor definitions of the main three independent scalar invariants (I, II and III) of rank-2 tensors. Express the main three independent scalar invariants (I, II and III) of rank-2 tensors in terms of the three subsidiary scalar invariants (I1 , I2 and I3 ). Referring to question 7.4, find the three scalar invariants (I, II and III) of D and the three scalar invariants (I1 , I2 and I3 ) of E using the tensor definitions of these invariants with systematic index substitution. State the following vector identities in tensor notation: ∇×r ∇ · (f A) A × (∇ × B) ∇ × (A × B) = = = = 0 f ∇ · A + A · ∇f (∇B) · A − A · ∇B (B · ∇) A + (∇ · B) A − (∇ · A) B − (A · ∇) B 7.9 State the divergence and Stokes theorems for a vector field in Cartesian coordinates using vector and tensor notations. Also, define all the symbols involved. 7.10 Prove the following vector identities using tensor notation and techniques with full justification of each step: ∇·r ∇ · (∇ × A) A · (B × C) ∇ × (∇ × A) = = = = n 0 C · (A × B) ∇ (∇ · A) − ∇2 A 7.4 Exercises and Revision 178 7.11 What is the type, in the form of (m, n, w), of the Riemann-Christoffel curvature tensor of the first and second kinds? 7.12 What are the other names used to label the Riemann-Christoffel curvature tensor of the first and second kinds? 7.13 What is the importance of the Riemann-Christoffel curvature tensor with regard to characterizing the space as flat or curved? 7.14 State the mathematical definition of the Riemann-Christoffel curvature tensor of either kinds in determinantal form. 7.15 How can we obtain the Riemann-Christoffel curvature tensor of the first kind from the second kind and vice versa? 7.16 Using the definition of the second order mixed covariant derivative of a vector field (see § 5.2) and the definition of the mixed Riemann-Christoffel curvature tensor, verify the following equation: Aj;kl −Aj;lk = Rijkl Ai . Repeat the question with the equation: Aj;kl − Aj;lk = Rjilk Ai . 7.17 Based on the equations in question 7.16, what is the necessary and sufficient condition for the covariant differential operators to become commutative? 7.18 State, mathematically, the anti-symmetric and block symmetric properties of the Riemann-Christoffel curvature tensor of the first kind in its four indices. 7.19 Based on the two anti-symmetric properties of the covariant Riemann-Christoffel curvature tensor, list all the forms of the components of the tensor that are identically zero (e.g. Riijk ). 7.20 Verify the block symmetric property and the two anti-symmetric properties of the covariant Riemann-Christoffel curvature tensor using its definition. 7.21 Repeat question 7.20 for the anti-symmetric property of the mixed Riemann-Christoffel curvature tensor in its last two indices. 7.22 Based on the block symmetric and anti-symmetric properties of the covariant RiemannChristoffel curvature tensor, find (with full justification) the number of distinct nonvanishing entries of the three main types of this tensor (see Eqs. 569-571). Hence, find the total number of the independent non-zero components of this tensor. 7.23 Use the formulae found in question 7.22 and other formulae given in the text to find the number of all components, the number of non-zero components, the number of zero components and the number of independent non-zero components of the covariant Riemann-Christoffel curvature tensor in 2D, 3D and 4D spaces. 7.24 Prove the following identity with full justification of each step of your proof: Raakl = 0. 7.25 Make a list of all the main properties of the Riemann-Christoffel curvature tensor (i.e. rank, type, symmetry, etc.). 7.26 Prove the following identity using the Bianchi identities: Rijkl;s + Riljk;s = Riksl;j + Rikjs;l . 7.27 Write the first Bianchi identity in its first and second kinds. 7.28 Verify the following form of the first Bianchi identity using the mathematical definition of the Riemann-Christoffel curvature tensor: Rijkl + Rkijl + Rjkil = 0. 7.29 What is the pattern of the indices in the second Bianchi identity? 7.30 Write the determinantal form of the Ricci curvature tensor of the first kind. 7.4 Exercises and Revision 179 7.31 Starting from the determinantal form of the Ricci curvature tensor of the first kind, obtain the following form of the Ricci curvature tensor with justification of each step √ a √ in your derivation: Rij = ∂j ∂i ln g + Γabj Γbia − √1g ∂a gΓij . 7.32 Verify the symmetry of the Ricci tensor of the first kind in its two indices. 7.33 What is the number of distinct entries of the Ricci curvature tensor of the first kind? 7.34 How can we obtain the Ricci curvature scalar from the covariant Riemann-Christoffel curvature tensor? Write an orderly list of all the required steps to do this conversion. 7.35 Make a list of all the main properties of the Ricci curvature tensor (rank, type, symmetry, etc.) and the Ricci curvature scalar. 7.36 Outline the importance of the Ricci curvature tensor and the Ricci curvature scalar in characterizing the space. 7.37 Write, in tensor notation, the mathematical expressions of the following tensors in Cartesian coordinates defining all the symbols involved: infinitesimal strain tensor, stress tensor, first and second displacement gradient tensors, Finger strain tensor, Cauchy strain tensor, velocity gradient tensor, rate of strain tensor and vorticity tensor. 7.38 Which of the tensors in question 7.37 are symmetric, anti-symmetric or neither? 7.39 Which of the tensors in question 7.37 are inverses of each other? 7.40 Which of the tensors in question 7.37 are derived from other tensors in that list? 7.41 What is the relation between the first and second displacement gradient tensors? 7.42 What is the relation between the velocity gradient tensor and the rate of strain tensor? 7.43 What is the relation between the velocity gradient tensor and the vorticity tensor? 7.44 What is the relation between the rate of strain tensor and the infinitesimal strain tensor? 7.45 What are the other names given to the following tensors: stress tensor, deformation gradient tensors, left Cauchy-Green deformation tensor, Cauchy strain tensor and rate of deformation tensor? 7.46 What is the physical significance of the following tensors: infinitesimal strain tensor, stress tensor, first and second displacement gradient tensors, Finger strain tensor, velocity gradient tensor, rate of strain tensor and vorticity tensor? References G.B. Arfken; H.J. Weber; F.E. Harris. Mathematical Methods for Physicists A Comprehensive Guide. Elsevier Academic Press, seventh edition, 2013. R.B. Bird; R.C. Armstrong; O. Hassager. Dynamics of Polymeric Liquids, volume 1. John Wiley & Sons, second edition, 1987. R.B. Bird; W.E. Stewart; E.N. Lightfoot. Transport Phenomena. John Wiley & Sons, second edition, 2002. M.L. Boas. Mathematical Methods in the Physical Sciences. John Wiley & Sons Inc., third edition, 2006. J. Bonet; R.D. Wood. Nonlinear Continuum Mechanics for Finite Element Analysis. Cambridge University Press, first edition, 1997. C.F. Chan Man Fong; D. De Kee; P.N. Kaloni. Advanced Mathematics for Engineering and Science. World Scientific Publishing Co. Pte. Ltd., first edition, 2003. T.L. Chow. Mathematical Methods for Physicists: A concise introduction. Cambridge University Press, first edition, 2003. P. Grinfeld. Introduction to Tensor Analysis and the Calculus of Moving Surfaces. Springer, first edition, 2013. J.H. Heinbockel. Introduction to Tensor Calculus and Continuum Mechanics. 1996. D.C. Kay. Schaum’s Outline of Theory and Problems of Tensor Calculus. McGraw-Hill, first edition, 1988. K.F. Riley; M.P. Hobson; S.J. Bence. Mathematical Methods for Physics and Engineering. Cambridge University Press, third edition, 2006. T. Sochi. Tensor Calculus Made Simple. CreateSpace, first edition, 2016. T. Sochi. Introduction to Differential Geometry of Space Curves and Surfaces. CreateSpace, first edition, 2017. I.S. Sokolnikoff. Tensor Analysis Theory and Applications. John Wiley & Sons, Inc., first edition, 1951. B. Spain. Tensor Calculus: A Concise Course. Dover Publications, third edition, 2003. J.L. Synge; A. Schild. Tensor Calculus. Dover Publications, 1978. D. Zwillinger, editor. CRC Standard Mathematical Tables and Formulae. CRC Press, 32nd edition, 2012. 180 Index Absolute covariant derivative, 130 derivative, 6, 11, 94, 109, 129–133, 136 differentiation, 94, 112, 129–132, 136 permutation tensor, 8, 81, 90, 99, 100, 107, 110 tensor, 57, 58, 72, 76, 81, 88, 89, 91, 107, 108, 167, 172 value, 11, 81 Active transformation, 33 Addition of tensors, 14, 62, 63, 73 Admissible transformation, 26, 31, 32, 34, 41, 47–49, 62, 112, 141 Affine coordinate system, 25, 30, 118, 167 tensor, 62, 112, 120, 132, 133 transformation, 30 Algebraic addition, 34, 73 subtraction, 73 Anisotropic tensor, 58, 72 Antisymmetric, 50, 59–62, 72, 73, 77, 78, 81, 84, 107, 169–171, 176, 178, 179 symmetric tensor, 59–62, 73, 176 symmetry, 59, 61, 73, 107, 169 Area, 6, 39, 76, 103–105, 110, 160 Associate metric tensor, 91 tensor, 94, 109 Associative, 63–65, 73 Associativity, 32, 63, 65 Asymmetric tensor, 59, 174 Axial tensor, 55 Basis tensor, 54, 71–73, 141 vector, 6, 7, 9, 15, 17, 24–28, 32, 39–46, 48, 49, 53–55, 69–72, 74, 76, 77, 91–94, 97, 99, 100, 109, 112–114, 120, 121, 123–125, 128– 130, 132, 134–136, 140, 141, 146, 148, 150, 152 Bianchi identity, 171, 172, 178 Block symmetric, 170, 178 symmetry, 169 Bound index, 14, 20 Cartesian, 1, 6, 7, 10, 15, 16, 22–30, 33–36, 39, 42, 47, 48, 51, 53–56, 66, 68, 69, 71, 72, 74, 83, 91–95, 98, 99, 108–111, 113, 118, 119, 121, 134, 137, 138, 140, 146, 152, 155, 161, 173–177, 179 Cauchy -Green deformation tensor, 175, 179 strain tensor, 6, 175, 179 stress law, 174 stress tensor, 174 Chain rule, 130, 131 Christoffel symbol, 6, 7, 94, 112–123, 128, 129, 132– 135, 144, 153, 171, 172 Circle, 25, 26, 35, 36 Closure, 32 Comma notation, 6, 10, 11, 19, 117 Commutative, 11, 32, 48, 63–65, 67, 73, 74, 85, 92, 124, 131, 135, 168, 178 Commutativity, 63, 84, 93, 126, 163–167 Complex, 9, 11 Composition of transformations, 32, 47 Conjugate metric tensor, 91 tensor, 94, 109 Conserved, 77, 78, 81, 107 Continuity condition, 11, 20, 31, 47, 124, 162, 169, 171, 172 Continuous, 11, 19, 22, 30, 31, 34, 127 Continuum mechanics, 69, 155, 173–175 Contracted epsilon identity, 87 Contraction, 14, 57, 62, 66–69, 73, 74, 82, 96, 124, 126, 132, 135, 137, 139, 144, 152 Contravariant, 7 basis vector, 6, 24–26, 39–41, 43–45, 48, 49, 53– 55, 69–71, 74, 91, 92, 109, 113, 132, 140 component, 43, 48, 49, 54, 69, 71, 97, 100, 102, 121, 144 differentiation, 121, 134 index, 51, 52, 58, 61, 93, 120, 122, 142–144, 168, 170, 172 Kronecker delta, 132 metric tensor, 7, 39, 43, 52, 70, 92, 93, 96, 109, 110, 114, 124, 125, 132, 142 partial derivative, 11 permutation tensor, 8, 84, 100, 107, 132 tensor, 32, 50–55, 65, 99, 121, 143, 153 tensor derivative, 11 181 Coordinate curve, 24–27, 35–40, 43–48, 53, 55, 103–106, 123 surface, 24–26, 35–40, 43–46, 48, 53, 55, 105, 110, 111 system, 7, 10, 12, 13, 15, 16, 20, 22–36, 39– 41, 44–51, 53, 55, 57, 58, 62, 70–72, 74–77, 83, 91–95, 97, 99, 102, 103, 105, 108–113, 118–121, 123–125, 128, 129, 132–138, 140, 144–148, 150, 152, 158, 161, 167, 171, 175 Covariant basis vector, 6, 24–27, 39–41, 43–45, 48, 49, 53–55, 69–71, 74, 91, 92, 99, 109, 113, 114, 132 component, 43, 48, 49, 54, 69, 71, 97, 100, 101, 121, 141, 143 derivative, 6, 11, 120–131, 133–136, 141–143, 145, 168, 178 differentiation, 10, 112, 120–132, 134–136 index, 51–53, 58, 93, 122, 137, 152, 170, 172, 173 Kronecker delta, 132 metric tensor, 7, 39, 42, 43, 52, 53, 70, 81, 92– 94, 96, 99, 103, 105, 108–110, 114, 124, 125, 132, 142, 172 permutation tensor, 8, 99, 132 Riemann-Christoffel curvature tensor, 167, 169, 171, 178 tensor, 32, 50–55, 64, 65, 67, 99, 110, 121, 145 Cross product, 43, 56, 57, 84, 99, 100, 103, 104, 110, 137, 139, 143, 152, 157, 163, 164, 177 Curl, 6, 56, 72, 139, 143, 144, 147, 149, 151–154 Curvature invariant, 173 scalar, 173 Curve, 6, 7, 11, 12, 23–25, 35, 36, 53, 76, 98, 99, 103, 110, 129–132, 136 Curved space, 22–24, 46, 94, 109, 110, 178 Curvilinear coordinate system, 24–26, 30, 34, 36, 40, 47, 48, 71, 120, 121, 123, 128, 133, 134, 137 Cylindrical coordinate system, 6, 8, 24–26, 28, 35, 37, 39, 47, 48, 71, 95, 110, 111, 113, 117– 119, 134, 137, 148, 150, 152, 153, 175 calculus, 11 geometry, 12, 17, 24, 119, 155 operator, 6, 10, 11, 51, 52, 121, 124, 126, 135, 137, 138, 140, 146, 152, 168, 178 Dimension of space, 6, 12, 15–18, 20, 24, 29, 30, 39, 44, 46, 58, 64, 65, 72, 76, 77, 82, 83, 107, 119, 156, 158, 174 Direct multiplication, 64–66 notation, 12 product, 56–58, 64, 65, 73 transformation, 29 Displacement gradient tensor, 6, 7, 174, 175, 179 vector, 6, 50, 102–106, 173 Distributive, 18, 64, 65, 67, 73 Distributivity, 163–165, 167 Divergence, 6, 137, 139, 142–145, 147, 149, 151–154 theorem, 160, 177 Dot product, 27, 42, 43, 53, 66–68, 74, 92, 93, 97, 98, 110, 114, 129, 130, 139, 156, 177 Double inner product, 68, 83, 84 Dummy index, 9, 14–18, 20, 52, 169 Dyad, 6, 55, 56, 65, 68, 72, 83, 122, 148–151 Dyadic multiplication, 64 product, 54 Ellipsoid, 23, 94 epsilon-delta identity, 87, 90, 108 Euclidean, 22–24, 27, 46, 94, 134, 167, 169 Exterior multiplication, 64 Extra pressure, 174 Finger strain tensor, 6, 175, 179 Flat metric, 94 space, 23, 27, 31, 46, 47, 94, 96, 109, 110, 127, 167, 168, 178 Fluid dynamics, 138, 173–175 mechanics, 69, 155, 173, 175 Form-invariance, 77 Form-invariant, 12, 13 Free index, 9, 14–18, 20, 51, 52, 59, 63, 66, 123, 161–167 Determinant, 6, 7, 11, 29, 30, 42, 43, 53, 72, 81, 85– 89, 91, 92, 94, 99, 102, 105, 108–110, 115, Gauss theorem, 155 126, 142, 156, 158, 172, 177 General Diagonal matrix, 44, 95 coordinate system, 1, 7, 10, 11, 20, 22, 24, 27, Differentiable, 31, 121–124, 127, 129–131, 134–136, 39, 41–43, 53, 57, 66, 70, 72–74, 91–93, 97, 138–144, 147, 149–152, 158, 160, 161 99, 100, 102, 103, 105, 106, 108–112, 119, Differential 182 120, 122, 133, 136, 137, 140, 141, 144–146, tensor, 58, 72, 76–78, 81 148, 150, 152, 153, 161, 167 Jacobian, 7, 29–33, 39, 42, 47, 48, 57, 94, 105 tensor, 62, 112, 120, 126, 133, 136, 143 matrix, 7, 29–31, 33, 42, 47 Generalized Kronecker delta, 7, 76, 82, 88–90, 108, 126 Kronecker delta, 7, 15, 41, 53, 58, 71, 76, 77, 82, Gibbs notation, 12 85–90, 93, 96, 107–109, 126, 127, 132, 135, Gradient, 6, 32, 39, 43, 45, 50, 129, 136–142, 144, 136 145, 147, 149, 150, 152–154, 175 vector, 32, 43 Laplacian, 6, 10, 137, 138, 140, 144–146, 148–154, 164 Handedness, 32, 55, 81, 105 Length, 6, 7, 15, 22, 23, 25, 27, 29, 39, 46, 53, 76, Homogeneous coordinate system, 23, 27, 29, 47, 110 83, 91, 92, 94, 96, 102, 103, 108, 110, 157 Hydrostatic pressure, 174 Levi-Civita identity, 87 Identity, 32 tensor, 58, 77 tensor, 22, 43, 76, 83, 91, 94 Linear transformation, 34, 48 algebra, 42, 44, 67, 74, 95, 155, 156 Imaginary, 9, 11, 27, 29, 92, 96 combination, 113 Improper coordinate system, 25 rotation, 58 operation, 124 transformation, 32–34, 48, 55, 77, 78, 81, 107 transformation, 26, 27, 29, 34, 62, 112, 121 Index Lorentz transformations, 23, 29, 95, 110 -free notation, 12, 13 Lowering operator, 52, 71, 93, 94, 97, 112–114, 126, notation, 9 168, 171–173 Indicial notation, 12, 13, 17, 19, 126, 135 structure, 17, 18, 20, 51, 52, 58, 63, 64 Infinitesimal, 6, 22, 23, 34, 46, 91, 103–106, 108, 111, 173 strain tensor, 7, 173, 175, 176, 179 Inner multiplication, 73, 74, 94, 113 product, 6, 57, 58, 64–69, 74, 83, 84, 94, 97, 98, 113, 124, 126, 127, 129, 131, 156 Integral, 105, 110 theorem, 155, 160, 161 Intrinsic derivative, 11, 130, 136 Invariance, 12, 16, 19, 71–73, 91, 112, 120 Invariant, 12, 13, 31, 32, 41, 47, 55, 58, 61, 62, 73, 76, 77, 98, 103, 121, 139, 141, 153, 155, 157, 158, 177 Inverse, 32 Jacobian, 7, 30, 31 of matrix, 44, 95, 156, 177 of metric, 44, 91, 92, 95, 109 of tensor, 174, 175, 179 transformation, 29–31, 47 Invertible matrix, 43, 91, 95, 158 transformation, 30, 31 Irrotational, 160 Isotropic, 23 Manifold, 22–24, 31, 39, 42, 46, 94 Matrix algebra, 65, 66, 74, 155, 156 notation, 156 Metric space, 22 tensor, 7, 15, 22, 23, 26, 27, 32, 39, 40, 42– 49, 52, 53, 69–71, 76, 81, 90–99, 103, 105, 107–110, 112–114, 118, 119, 123–126, 129, 131–136, 142, 147, 148, 150, 167–169, 171, 172 Minkowski metric, 95, 110 space, 23, 29 Mixed derivative, 10, 11, 127, 128, 135, 168, 169, 178 Kronecker delta, 7, 77, 82, 93, 132 metric tensor, 7, 22, 41, 43, 91–93, 109, 124, 125, 132 Riemann-Christoffel curvature tensor, 168, 169, 178 tensor, 7, 14, 18–22, 32, 41, 43, 51, 52, 54, 63, 65, 66, 69, 71, 77, 81, 82, 91–94, 102, 109, 121, 122, 124, 125, 131, 132, 141, 143, 152, 168, 169, 178 Multiplication by scalar, 62, 64, 73 183 of matrices, 67, 73, 92, 156, 177 product, 6, 64–68, 94, 124, 126, 131, 138 of matrix by vector, 67, 156 of tensors, 14, 58, 62, 64–67, 73, 74, 89, 94, 113, Parallelepiped, 41, 105 Partial 124, 131 derivative, 6, 10, 11, 30, 31, 53, 96, 112–114, Mutually 117, 118, 120–124, 126–130, 132, 133, 135, exclusive, 50 168, 169 independent, 24, 40, 44 differential operator, 11, 51, 52, 121, 138 orthogonal, 15, 25, 26, 36, 39, 43, 44, 49, 53, differentiation, 10, 83, 96, 120, 121, 124, 134, 83, 92 135 perpendicular, 24–27, 55, 94 Passive transformation, 33 nabla operator, 6, 137–140, 143, 146, 148, 150, 152, Perimeter, 161 Permissible transformation, 31, 62, 94, 98 153 Permutation Negative orthogonal transformation, 34, 48 of tensor, 68, 69, 74 Nontensor, 8, 57, 76–78, 80–90, 99, 100, 102, 107, Euclidean, 167 110, 132, 139 scalar tensor, 24, 32, 64, 65, 112, 120, 121, 128, Perpendicular, 6, 24–27, 39, 53, 55, 94 133, 134, 136, 146 Physical singular, 91, 92 basis vector, 71, 74 Nonlinear transformation, 26, 34 component, 69–71, 75, 137, 147, 148, 150, 153 Normal dimension, 25, 46, 69 stress, 174 representation, 6, 39, 69, 71, 74 vector, 7, 174 Plane, 23, 25, 26, 35, 36, 46, 55, 94 Normalized vector, 6, 102 Polar Oblique coordinate system, 25, 26 angle, 36 Order coordinate system, 8, 36, 148 of derivative, 10, 11, 127, 128, 135, 168, 169, tensor, 55 178 Polyad, 55 of indices, 10, 16–21, 51, 52, 55, 64, 66, 83, 93, Position vector, 6, 7, 39, 53, 102, 146, 158, 161 94, 109, 119, 123, 134, 141, 162 Positive orthogonal transformation, 33, 34, 48 of multiplicands, 67, 73, 124, 131, 157, 162–164 Principle of invariance, 12, 19, 112 of operators, 11, 18, 138, 162 Product rule, 96, 114, 121–125, 130, 132, 141, 142, of tensor, 14, 16, 18, 20 145, 162, 163, 165, 166 of transformations, 32 Proper Orthogonal, 15, 25–27, 36, 39, 43, 44, 49, 53, 83, 92 rotation, 58 coordinate system, 7, 24–26, 36, 39, 41, 43, 47, tensor, 55 70, 71, 74, 94, 95, 103–105, 108–111, 115– transformation, 32–34, 48, 77, 78, 81, 107 117, 119, 133, 134, 137, 146–148, 150, 153 Pseudo transformation, 27, 33, 34, 48, 55, 78, 81 tensor, 55–58, 63, 72, 78, 81 Orthonormal vector, 55, 57, 72, 157 basis set, 15, 43, 49, 53, 71, 83, 148, 150 Cartesian, 15, 16, 22, 23, 27, 28, 30, 39, 48, 51, Quotient rule of differentiation, 69 55, 66, 68, 71, 74, 83, 91, 93–95, 98, 99, tensor, 62, 69, 74 108–111, 118, 119, 134, 177 coordinate system, 15, 16, 20, 22, 23, 27, 28, Radius, 23 30, 39, 48, 51, 55, 66, 68, 71, 74, 83, 91, Raising operator, 52, 71, 93, 94, 97, 112, 121, 126, 93–95, 98, 99, 108–111, 118, 119, 134, 177 141, 142, 153, 168, 171–173 vectors, 6, 15, 20, 49, 72, 83, 84, 99, 108 Range of index, 12–16, 18, 65, 79, 80, 160 Orthonormalized vectors, 6, 53, 72 Rank Outer -0 tensor, 9, 12, 13, 16, 19, 58, 62, 65, 67, 153, multiplication, 64, 67, 73 157 184 -1 tensor, 9, 12–14, 19, 50, 56, 58, 64, 67, 70, 78, 140, 152, 157 -2 tensor, 9, 10, 12–14, 16, 19, 22, 54, 56, 58– 60, 64–68, 70–74, 76, 78, 79, 85, 86, 91, 97, 102, 107, 110, 121, 122, 134, 136–139, 141, 143, 149, 151–153, 156–158, 160, 161, 172–177 -3 tensor, 12–14, 19, 54, 60, 66, 72, 77–84, 86– 88, 99, 107, 112 -4 tensor, 65, 66, 72, 79, 107, 167 of tensor, 9, 12, 14, 16, 19, 20, 42, 50–54, 58, 59, 61–70, 72, 73, 77–79, 84, 88, 93, 97, 107, 120, 123, 128, 130–132, 135, 137–143, 146, 152, 157, 160, 161 Rate of deformation tensor, 176, 179 strain tensor, 7, 176, 179 Real, 9, 11, 24, 27, 29, 35, 44, 103 Reciprocal, 30, 31, 39, 41, 43, 44, 53, 63, 91, 92, 95 metric tensor, 91 Reciprocity relation, 53, 54, 72, 91 Rectangular Cartesian, 25, 26, 29, 33, 34, 54, 56, 72, 92 coordinate system, 25, 26, 29, 33, 34, 54, 56, 62, 72, 92, 112, 129, 132, 146 parallelepiped, 105 Rectilinear coordinate system, 24–26, 30, 34, 35, 47, 121, 123–125, 129, 132, 136 Reference frame, 32 Reflection, 33, 34, 55 Relative permutation tensor, 8, 81, 99, 107, 110 scalar, 94, 128, 136 tensor, 7, 57, 58, 63, 66, 72, 78, 81, 89, 107, 128, 136 Replacement operator, 52, 82, 88, 93, 107, 126, 136 Rheology, 175 Ricci curvature scalar, 7, 172, 173, 179 curvature tensor, 7, 172, 173, 178, 179 theorem, 114, 115, 124–126, 135 Riemann-Christoffel curvature tensor, 7, 22, 23, 127, 128, 167–172, 178, 179 Riemannian curvature, 23 geometry, 22, 24, 46 metric, 22, 94 space, 22, 23, 46, 93, 170 Right handed system, 41, 42, 49, 84 Rotation, 32–34, 58, 175, 176 Scalar, 6, 9–13, 16–20, 31, 34, 58, 59, 62–66, 68–70, 72, 73, 94, 103, 123, 124, 128–132, 135– 145, 147–151, 153, 157 field, 19, 50, 129, 136, 152–154, 158 invariant, 155, 157, 158, 177 multiplication, 34 operator, 137, 138, 152 triple product, 41, 84, 100, 101, 105, 106, 110, 111, 157, 177 Scale factor, 7, 36, 39, 41, 48, 69–71, 95, 100, 103, 105, 110, 111, 117–119, 146, 148, 150, 153 Schur theorem, 23, 46 Semicircle, 26, 36 plane, 26, 35 Semicolon notation, 6, 10, 11, 19, 124 Shear stress, 174 Shifting operator, 40, 52, 93, 98, 109, 113, 114, 126, 135, 136 Skew-symmetric, 59, 73, 169 Solenoidal, 160 Sphere, 23, 25, 26, 35, 36, 46 Spherical coordinate system, 6, 7, 10, 24–26, 28, 35, 38, 39, 47, 48, 71, 75, 95, 110, 111, 113, 117–119, 134, 137, 150, 152–154, 175 Stokes theorem, 155, 160, 161, 177 Straight line, 23, 25, 26, 35, 36, 47, 55 Stress tensor, 8, 174, 179 Subtraction of tensors, 62, 63, 73 Sum rule, 132 Summation, 11, 14–16, 20, 43, 62, 66, 113, 147 convention, 11, 14, 15, 19, 83, 87, 137, 148, 150 Surface, 6, 7, 12, 23–25, 35, 46, 53, 76, 94, 103–105, 119, 160, 161, 174 Symbolic notation, 9, 12, 19, 65, 67, 124, 126, 131, 135, 139, 156 Symmetric, 18, 22, 50, 59–62, 72, 73, 76, 77, 84, 91–93, 109, 113, 116, 118, 133, 169–172, 174–176, 178, 179 tensor, 59–62, 73, 84, 172, 176 Symmetry, 59, 61, 62, 73, 84, 107, 108, 119, 144, 153, 169, 173, 178, 179 Tangent vector, 32, 39, 43, 55, 98, 103, 129, 131 Tensor algebra, 42, 62, 63, 156 calculus, 1, 11, 12, 14, 17, 19, 22, 24, 57, 62, 65, 67, 69, 75–77, 86, 91, 112, 121, 137, 152, 155, 157, 161, 167, 168, 173, 177 component, 18, 20, 55, 78, 86, 112, 123, 132, 143, 152 equality, 14–18, 20, 58, 62, 63, 72, 73, 77 185 expression, 14–18, 20, 56, 58, 63, 72, 87 field, 19, 21, 136, 152, 160, 161 identity, 76, 87, 89, 155, 161 multiplication, 64, 65, 73, 124, 131 notation, 12, 13, 17, 71, 84, 92, 99, 109, 110, 137, 152, 153, 155, 157, 158, 160, 161, 173– 177, 179 representation, 50, 69–71, 74 term, 14–18, 20, 56–58, 60, 62, 63, 72, 82, 84, 86, 120, 122–124, 126, 128, 132, 134–136, 144, 162, 164, 172 test, 69, 74 Torus, 23 Total derivative, 112, 130, 132, 133, 136 differentiation, 112, 130, 132, 136 stress tensor, 174 Totally anti-symmetric, 61, 72, 78, 81, 84, 107 covariant Riemann-Christoffel curvature tensor, 167 symmetric, 61, 72 Trace, 7, 66, 83, 139, 156, 158, 177 Traction vector, 7, 174 Transformation, 7, 11–13, 15, 19, 22, 26, 27, 29–34, 36, 40–42, 47–51, 55, 57, 58, 61, 62, 69, 71–74, 76–78, 81, 91, 93, 94, 98, 103, 105, 107, 109, 112, 121, 141, 153, 157 Translation, 33, 34, 175 Triad, 54, 65 True scalar, 31, 157 tensor, 55–58, 63, 72 vector, 31, 55, 57, 72 notation, 12, 158, 160, 177 operator, 137, 138, 140, 152 triple product, 101, 102, 110, 157, 177 Velocity gradient tensor, 6, 175, 176, 179 vector, 7, 175 Volume, 6, 7, 39, 41, 105, 106, 111, 160 Vorticity tensor, 7, 8, 176, 179 Weight of tensor, 7, 53, 57, 58, 63, 64, 66, 72, 73, 78, 81, 89, 94, 107, 128, 136 Zero Unit tensor, 14 vector, 55, 56, 84, 160 Unity matrix, 53 tensor, 14, 92, 94, 109, 175 Vector, 6, 7, 9, 11–13, 15–17, 19, 20, 25–27, 31, 32, 39–44, 48, 53–59, 64–74, 83, 84, 87, 91–94, 97–103, 105, 108, 110, 111, 113, 121–124, 127–131, 134, 137–144, 146, 147, 149–153, 156–158, 160, 163, 164, 168, 174, 175, 177 algebra, 155 calculus, 74, 137, 140, 147, 148, 152, 155, 157, 158 field, 19, 138, 152–154, 158, 160, 177, 178 identity, 87, 161, 177 186 tensor, 14, 58, 62, 72, 73, 133 vector, 58 Author Notes • All copyrights of this book are held by the author. • This book, like any other academic document, is protected by the terms and conditions of the universally recognized intellectual property rights. Hence, any quotation or use of any part of the book should be acknowledged and cited according to the scholarly approved traditions. 187 About the Book This book is an intermediate level text on tensor calculus. The formulation is largely based on assuming a general curvilinear coordinate system. The book is furnished with an index and thorough sets of exercises. Cross referencing, which is hyperlinked in the electronic version, is used extensively throughout the book to facilitate linking related topics. The book also contains a number of graphic illustrations to visualize the abstract ideas and consolidate the subtle concepts. The book can be used as a textbook or as a reference for an introductory or intermediate level course on tensor calculus or related mathematical and scientific subjects. About the Author The author of the book possesses a diverse academic and research background. He holds a BEng in electronics engineering, a BSc in physics, a PhD in petroleum engineering, a PhD in crystallography, and a PhD in atomic physics and astronomy. He also worked as a postdoctoral research associate for several years. He published dozens of scientific research papers in many refereed journals and produced numerous academically oriented documents which are freely available on the world wide web.