Balakrishnan PDF
Balakrishnan PDF
Balakrishnan PDF
Note: All figures have been omitted in this write-up. The reader will find it an instruc-
tive exercise to fill in all necessary figures.
Table of contents
1
6. Multivalued functions; integral representations: Branch points and branch
cuts. Algebraic and logarithmic branch points, winding point. Contour integrals
in the presence of branch points. An integral involving a class of rational func-
tions. Contour integral representations for the gamma, beta, and Riemann zeta
functions. Connection with Bernoulli numbers. The statement of the Riemann
hypothesis. Contour integral representations of the Legendre functions Pν (z) and
Qν (z). Singularities of functions defined by integrals. End-point and pinch singu-
larities, examples. Singularities of the Legendre functions. Dispersion relations
for the Legendre functions.
9. Quiz 1; solutions.
10. The fundamental Green function for ∇2 : Green functions. Poisson’s equa-
tion; the fundamental Green function for ∇2 . Solution for a spherically symmet-
ric source. The Coulomb potential in d dimensions. Ultraspherical coordinates.
A divergence problem. Dimensional regularization. A direct derivation using
Gauss’ Theorem. The Coulomb potential in d = 2 dimensions. A direct deriva-
tion once again.
11. The diffusion equation: Fick’s laws of diffusion. The fundamental solution
in d dimensions. Solution for an arbitrary initial distribution. Moments of the
distance travelled in time t. Diffusion in one dimension: continuum limit of a ran-
dom walk. The Smoluchowski equation. Sedimentation. Absorbing and reflecting
boundary conditions. Free diffusion on a semi-infinite line. Finite boundaries:
solution by the method of images. Solution by separation of variables. Survival
probability. First-passage-time distribution and mean first-passage time. Con-
nection with the Schrödinger equation for a free particle. Spreading of a quantum
mechanical wave packet.
2
12. The Green function for (∇2 +k 2 ); nonrelativistic scattering: The Helmholtz
operator. The scattering amplitude; differential and total cross-sections. Inte-
gral equation for scattering. Green function for the Helmholtz operator. Exact
formula for the scattering amplitude. Scattering geometry and the momentum
transfer. Born series and the Born approximation. The Yukawa and Coulomb
potentials. The Rutherford scattering formula.
13. The wave equation: Formal solution for the causal Green function. The
solution in (1 + 1), (2 + 1) and (3 + 1) dimensions. Retarded solution of the wave
equation. Remarks on propagation in spatial dimensions d > 3.
14. The rotation group and all that: Rotations of the coordinate axes. Orthog-
onality of rotation matrices. Proper and improper rotations. Generators of
infinitesimal rotations in 3 dimensions. The general rotation matrix in 3 dimen-
sions. The finite rotation formula for a vector. Relation between the groups
SO(3) and SU(2). Rotation generators in 3 dimensions transform like a vector.
Derivation of Hadamard’s lemma. The general form of the elements of U(2) and
SU(2). The parameter spaces of SU(2) and SO(3). Tensor and spinor represen-
tations. The universal covering group of a Lie group. The group SO(2) and its
covering group. The groups SO(n) and Spin (n). Parameter spaces of U(n) and
SU(n). A bit about the first homotopy group of a space.
3
1 Analytic functions of a complex variable
We begin this course with the study of analytic functions of a complex variable. The
theory of functions of a complex variable (and its generalization, the theory of functions
of several complex variables) is one of the richest and most beautiful branches of math-
ematics, with deep results and far-reaching implications and applications. It is also a
vast subject. Only its most rudimentary aspects will be dealt with in what follows.
No formal proofs of theorems will be given, as usual, and the treatment of standard
material will be heuristic rather than rigorous. I shall assume that you already have
some familiarity with analytic functions of a complex variable, from an earlier course
on mathematical methods in physics.
1. Show that the real and imaginary parts of the complex numbers listed below are as
indicated. Remember the standard phase convention, according to which i = eiπ/2 and
−i = e−iπ/2 = e3πi/2 .
(a) (i)i = cos 12 πe−π/2 + i sin 12 πe−π/2 .
i
∞
(iπ)2n+1
(b) = 0.
n=0
(2n + 1)!
n
π π π π
∞
1 1+i
(c) √ = cosh cos cos sin + i sinh cos sinh sin .
n=0
(2n)! 2 8 8 8 8
4
(e) The infinite continued fraction
√ √
i i i 1 17 + 1 17 − 1
· · · ad inf. = − + √ +i √
1+ 1+ 1+ 2 2 2 2 2
0.3002 + 0.6248 i.
Hint: In (d), (e) and (f), calling the left-hand side z, you have to solve the respec-
tive equations z = (i + z)1/2 , z = i/(1 + z) and z = 1/(i + z). Each of these reduces
to a quadratic equation, and you must take care to identify the correct root in each case.
Equations to curves in the plane in terms of z and z ∗ : We know that the Carte-
sian coordinates x and y labeling points on a plane are linearly independent of each
other. It is important to recognize that, equivalently, the complex variable z = x + iy
and its complex conjugate z ∗ = x − iy are linearly independent. Once you bear this
in mind, it becomes much easier to understand the concept of analytic functions of a
complex variable z.
2. Equations to curves in the xy-plane are often very conveniently expressed in terms
of z and z ∗ . Here are some examples of familiar curves, thus expressed:
(a) In terms of the complex variable z, write down the equation of the ellipse in the
xy-plane whose foci are at the points ±x0 on the x-axis, and whose semi-major
axis is equal to a (where a > x0 ).
(b) Find the locus of points in the z-plane given by the following conditions :
2
(i) |z − 1| − |z + 1| = 1 (ii) (z − 1)/(z − 2) = 1 (iii) ez = e4 .
The Riemann sphere and stereographic projection: We know that the number
line extends from −∞ on the left to +∞ on the right. In the complex plane, z may tend
to infinity along any of an infinite number of directions. It is convenient to ‘compactify’
the plane by bringing together all these points at infinity and ‘glueing’ them together
into a single point. This can be done by a mapping between the complex plane and
the surface of a sphere. Consider a sphere of unit radius with its center at the origin
5
of coordinates. The coordinates of any point on the surface are given by (ξ1 , ξ2 , ξ3 ),
where
ξ12 + ξ22 + ξ32 = 1.
In terms of spherical polar coordinates on the unit sphere, we have of course
ξ1 = sin θ cos ϕ , ξ2 = sin θ sin ϕ , ξ3 = cos θ .
The coordinates of the ‘north pole’ N are (0, 0, 1), while those of the ‘south pole’ S are
(0, 0, −1).
Now consider the equatorial plane of the sphere. Let the Cartesian coordinates
on this plane be (x , y), with the same origin as the sphere, and the x and y axes
running along the ξ1 and ξ2 axes, respectively. The stereographic projection of any
point P on the sphere onto the equatorial plane is obtained by joining N and P by a
straight line. The projection of P is the point P where the line cuts the equatorial
plane. This plane is regarded as the complex plane, while the sphere is called the
Riemann sphere. The coordinates x and y of P in the complex plane are related to
the coordinates (ξ1 , ξ2 , ξ3 ) on the Riemann sphere by
ξ1 ξ2
x= = cot 12 θ cos ϕ and y = = cot 12 θ sin ϕ.
1 − ξ3 1 − ξ3
Therefore
ξ1 + iξ2
z = x + iy = = cot 12 θ eiϕ .
1 − ξ3
It is evident that points of the northern hemisphere are mapped to points outside the
unit circle in the complex plane; points in the southern hemisphere are mapped to
points inside the unit circle; and points on the equator are mapped to themselves as
points on the unit circle |z| = 1.
As P gets closer to the point of projection N, it is clear that the point P moves
farther and farther away from the origin in the complex plane. The point at infinity
in the complex plane is defined as the image of the north pole under the projection,
and denoted by ∞ . The finite part of the complex plane, together with the point at in-
finity, is called the extended complex plane. In what follows, by the term ‘complex
plane’ we shall generally mean the extended complex plane. It is very advantageous
to have ‘infinity’ identified with a single point in this fashion, and to be able to specify
∞ as a specific value for the complex number z.
6
(b) What is the projection in the complex plane of a circle of latitude on the Riemann
sphere? Of a meridian of longitude?
• In general, circles on the Riemann sphere are mapped into circles or straight lines
in the complex plane by stereographic projection.
(c) Let P and Q be any two points on the Riemann sphere, whose projections on the
complex plane are given by z1 and z2 . Show that the chordal distance between
P and Q (i.e., the length of a chord drawn from P to Q) is given by
2 |z1 − z2 |
d(z1 , z2 ) = .
(|z1 |2 + 1)(|z2 |2 + 1)
With this definition of the distance, the exceptional role played by the point at infinity
causes no trouble, as it does in the case of the ordinary Euclidean definition |z1 − z2 | .
We have
2
d(z, ∞) = .
|z|2 + 1
Clearly, d(z1 , z2 ) ≤ 2 for any two points z1 and z2 in the extended complex plane.
7
• Hence an entire function that is not just a constant cannot be analytic at the
point at infinity.1
One way to understand what is meant by an analytic function of z is as follows.
Recall that, just as x and y are linearly independent coordinates on the plane, so are
the linear combinations z = x + iy and z ∗ = x − iy. Hence any function of x and y can
equally well be written as a function of z and z ∗ . An analytic function of z is then a
sufficiently well-behaved function2 of x and y that depends on the combination x + iy
alone, and that does not involve the other combination, namely, x − iy. Thus, f is an
analytic function of z if ∂f /∂z ∗ ≡ 0. But, by the chain rule of differentiation,
∂f ∂f ∂f
=0 =⇒ −i = 0.
∂z ∗ ∂x ∂y
The real and imaginary parts of the last equation yield precisely the Cauchy-Riemann
conditions. These conditions may therefore be written compactly as ∂f /∂z ∗ ≡ 0.
Writing any function of x and y in terms of z and z ∗ enables us to identify functions
that cannot be analytic functions of z, merely by checking whether z ∗ makes its ap-
pearance in the function. If it does so, the function cannot be an analytic function of z.
4. Identify which among the following are analytic functions of z in some region, and
which are not. In the former case, identify the region of analyticity.
(a) x (b) ix − y (c) r
(d) eiθ
(e) x − iy
2 2
(f) x − iy
(g) (x + iy) 2
(h) x − y − 2ixy
2 2
(i) i tan−1 (y/x)
(j) (iy + x2 + y 2) (k) [(x + i)2 − y 2 ]1/2 (l) (x − iy)/(x2 + y 2 )
(m) x4 + 2ix2 y 2 − y 4 (n) i ex sin y (o) x2 + x + 1 − y 2 + iy (2x + 1).
The real and imaginary parts of an analytic function : Let u and v be the real
and imaginary parts of an analytic function f (z). It follows from the Cauchy-Riemann
conditions that both u and v individually satisfy Laplace’s equation in two dimensions,
i.e., ∇2 u = 0 and ∇2 v = 0.
• The real and imaginary parts of an analytic function are thus harmonic functions.
The combination u + iv constitutes an analytic function in the intersection of the
regions in which they are individually harmonic.
• It is also evident from the Cauchy-Riemann conditions that an analytic function
cannot be identically equal to a purely real or purely imaginary function, except
in the trivial case when the function is just a constant.
1
In other words, it must be singular at z = ∞. We will discuss singularities subsequently.
2
By the phrase ‘sufficiently well-behaved’, we mean that the real and imaginary parts of the
function have continuous partial derivatives with respect to x and y.
8
• An analytic function that is real when its argument is real, i.e., a function such
that f (x) = u(x, 0) is real (or v(x, 0) vanishes identically), is called a real ana-
lytic function.
• It follows from the Cauchy-Riemann conditions that the scalar product (∇u ·
∇v) = 0. The curves u = constant and v = constant thus constitute two mutu-
ally orthogonal families of curves in the complex plane, for any analytic function.
5. Given u (or v) as indicated below, find the corresponding v (respectively, u), either
by inspection or by integrating the Cauchy-Riemann conditions. Indicate also the
region of analyticity of f (z) in each case.
9
have continuous partial derivatives, the definition of the derivative given above yields,
in the limit → 0,
df (z) −iα ∂u ∂v ∂v ∂u
=e cos α + i sin α + i cos α − i sin α .
dz ∂x ∂y ∂x ∂y
The remarkable fact is that this expression becomes completely independent of α if
and only if
∂u ∂v ∂u ∂v
= and =− ,
∂x ∂y ∂y ∂x
which are precisely the Cauchy-Riemann conditions.
• Analyticity may therefore be understood as equivalent to the requirement that
the limit defining the derivative of a complex function be unique, independent of
the direction along which z + δz approaches z in the complex plane.
7. In its region of analyticity, and analytic function f (z) may also be regarded as a
map from (a region of) the complex plane to (a region of) the complex plane.
(a) Show that (the determinant of) the Jacobian of the transformation (x , y)
(u , v) is just |f (z)|2 , where f (z) denotes the derivative of f (z).
(b) Show that ∇2 (|f (z)|2 ) = 4 |f (z)|2 .
If f (z) and z0 are specified, the coefficients an in the Taylor series above are uniquely
determined. We have
1 dn f (z)
an = ,
n! dz n z=z0
10
the existence of derivatives of all orders being guaranteed by the analyticity of f (z).
The radius of convergence itself is determined by the coefficients {an } —more precisely,
by the asymptotic (or n → ∞) behavior of an . The radius of convergence is given by
an
R = lim ,
n→∞ an+1
where lim sup (or supremum) stands for the least upper bound of the quantity con-
cerned. It is important to keep in mind that
∞
zn ∞
(z − 3)n ∞
(z/2)n
∞
zn
(a) (b) √ (c) (d)
n=0
(n + 1)3 n=0 n! n=1
n(n + 1) n=1
ln (n + 1)
∞
zn
∞
(z − 2)2n
∞
zn
∞
2
(e) (f) (g) (h) zn
n=0
(n + 1)n n=0
(2n)! n=1
ln ln (n + 1) n=0
√
∞
∞
zn ∞
(z − i 2)n
∞
n!
(i) z (j) (k) (l) n7 z n
n=0 n=0
(n!)1/6 n=0
n! n=0
∞ ∞ ∞ ∞
n
2n (ln n) z n (ln ln n) z n 1−z
(m) (z) (n) (o) (p)
n=0 n=1
n n=2
ln n n=0
1+z
∞ ∞ 2 n ∞ ∞
(2z)n z −1 (z/4)n Γ(n − 12 ) n
(q) (r) (s) (t) z .
n=0
n2 + 1 n=0
z2 + 1 n=0
n+2 n=1
Γ(n)
If R is infinite for a power series centered about any finite point z0 , this means
that the function represented by the power series is an entire function. A power series
diverges outside its circle of convergence. On the circle of convergence itself, its behavior
can be quite complicated, and change from one point on the circle to another. The
series may diverge at one or more points on the circle; or it may converge at some
11
points, diverge at others, and oscillate at others; it may even converge at all points.
What is guaranteed, however, is that the function that the power series represents
inside the circle of convergence has at least one singularity on that circle. Remarkably
enough, this remains true even if the power series itself converges absolutely at all
points on the
∞circle of convergence. An example of this phenomenon is provided by
the series n=1 z /n2 . This series converges absolutely at all points on its circle of
n
convergence, which is just the unit circle |z| = 1. (It attains its largest value on the
circle at the point z = 1, where the sum is equal to ζ(2) = π 2 /6.) And yet, the function
represented by the series is singular at z = 1. At the other extreme, we have cases in
which the function represented by the series has a dense setof singularities
∞everywhere
∞ n
on the circle of convergence. Examples of such series are n=0 z and n=1 (z)2 , for
n!
each of which the unit circle |z| = 1 is the circle of convergence. We shall return to
some of these aspects later on.
12
2 Calculus of residues
Cauchy’s integral theorem : As you know, the property of analyticity of a function
f (z) implies that the function is infinitely differentiable, and that each derivative is also
an analytic function. What does analyticity imply for the integral of a function? Once
again, the Cauchy-Riemann condition can be used to arrive at the following conclusion:
Suppose f (z) is analytic in some region R. Then, if z1 and z2 are two points lying in
R, the line integral z2
dz f (z)
z1
of the function is independent of the actual path between the end-points, provided of
course that the path lies entirely in R. The value of the line integral is dependent only
on the end points.3 As a consequence, the path connecting the end-point z1 and z2 can
be distorted like a flexible and stretchable piece of string, and the value of the integral
does not change.
13
is not defined at z = 0, and appears to have a singularity at that point. However,
limz→0 f (z) = 1, and we may define f (0) = 1 so as to preserve continuity. This gets
rid of the ‘removable’ singularity and ensures the continuity of f (z) at z = 0. In
all that follows, it will be assumed that functions are defined at such removable sin-
gularities in accord with the limiting values concerned, so that continuity is guaranteed.
Simple pole: The simplest nontrivial singularity is a simple pole. f (z) has a simple
pole at the point z = a if, in a neighborhood of that point, it can be expressed in the
form
c−1 ∞
f (z) = + cn (z − a)n .
(z − a)
n=0
singular part regular part
The first term on the right is called the ‘singular part’ of f (z) at z = a. It becomes
infinite at z = a. The coefficient c−1 is a non-zero, finite constant (in general, a com-
plex number) that is called the residue of f (z) at z = a. The notation c−1 is used to
remind us that it is the coefficient of (z − a)−1 in the expansion of f (z) in powers of
(z − a). The ‘regular part’ stands for a function that is analytic in the neighborhood
concerned, and hence is expandable in a convergent power series in (non-negative)
powers of (z − a), in the neighborhood concerned. In some cases the series may, of
course, terminate after a finite number of terms.
The crucial point is that, at a simple pole, the singular part comprises just the
single term c1 (z − a)−1 . This term may or may not be explicit. Consider, for example,
the function (sin z)/z 2 . In the neighborhood of the origin, it has the expansion
sin z 1 (−1)n z 2n+1
∞
1 z2 z4
= = + − + − . . . .
z2 z 2 n=0 (2n + 1)! z
3! 5!
sing. part reg. part
This function therefore has a simple pole at z = 0, with residue equal to 1. Incidentally,
it has no other singularities in the finite part of the complex plane (remember that sin z
is an entire function). As a second example, consider the function cosec z = 1/(sin z).
Since sin z has simple zeroes at all integer multiples of π, i.e., at z = nπ where n ∈ Z,
it follows that cosec z has simple poles at those points. Using the fact that sin nπ = 0
and cos nπ = (−1)n , the Taylor expansion of sin z near z = nπ is given by
sin z = (−1)n (z − nπ) − 16 (−1)n (z − nπ)3 + . . . .
Hence
(−1)n (−1)n
cosec z = + 16 (−1)n (z − nπ) + . . . = + regular part.
(z − nπ) (z − nπ)
Thus the residue of cosec z at its simple pole at z = nπ is (−1)n .
14
In general, if f (z) has a simple pole at z = a, then its residue at the pole is given
by
c−1 = Res f (z) = lim [(z − a) f (z)].
z=a z→a
If f (z) is of the form g(z)/h(z) where g(z) and h(z) are analytic at z = a, and h(z)
has a simple zero at z = a while g(a)
= 0, then f (z) has a simple pole at z = a. We
have
g(z) g(a)
f (z) = =⇒ Res f (z) = ,
h(z) z=a h (a)
where the prime denotes the derivative of the function concerned.
(c) sech z has a simple pole at z = i(n + 12 )π, with residue i (−1)n+1 .
(e) tanh z has a simple pole at z = i(n + 12 )π, with residue +1.
Note, in particular, how the residue becomes independent of n in the last four cases.
This property will be useful when we deal with the summation of certain series.
Multiple pole: The next case is that of a pole of higher order, or a multiple pole.
The function f (z) has a pole of order m at z = a if, in the neighborhood of that point,
it can be expressed in the form
c−m c−1
∞
f (z) = +···+ + cn (z − a)n ,
(z − a)m (z − a) n=0
singular part regular part
where m is a positive integer. The residue of f (z) at the multiple pole is still the
coefficient c−1 . The significance of the residue will become clear shortly, when we
consider Cauchy’s integral theorem. For the present, note that the residue at a multiple
pole of order m is given by
dm−1
1
c−1 = Res f (z) = lim [(z − a)m f (z)].
z=a z→a (m − 1)! dz m−1
15
This formula is a generalization of that for the residue at a simple pole.
Essential singularity; Laurent series: Going a step further, the singular part may
involve all negative integral powers of (z − a). The function f (z) has an (isolated)
essential singularity at z = a if, in the neighborhood of that point (see below), it
can be expressed in the form
∞
c−n
∞
f (z) = + cn (z − a)n .
(z − a) n
n=1 n=0
singular part regular part
Once again, the coefficient c−1 is the residue of f (z) at the singularity. A representation
such as the foregoing in positive as well as negative powers of (z−a) is called a Laurent
series. The question now arises as to the region in which such a representation is valid.
The regular part, as we know, is convergent inside some circle of convergence centered
at z = a, and of radius r1 , say. By setting w = (z − a)−1 , it follows that the singular
part is convergent inside some circle of convergence in the complex w-plane that is
centered at the origin. In other words, it is convergent outside some circle centered at
z = a, of radius r2 , say. Then, provided r2 < r1 , there is an annular overlap region
of inner radius r2 and outer radius r1 , in which both the infinite series are absolutely
convergent. This annulus is the region in which the Laurent series is valid. It may so
happen, of course, that r2 = 0. Then the Laurent series is valid in the punctured disc
of radius r1 centered at z = a, with only the point z = a left out. For example, the
function
∞
1
1/z
e = +
1
n! z n
n=1 reg. part
sing. part
16
— The function ez has an essential singularity at z = ∞. Hence each of the functions
cos z, sin z, cosh z and sinh z also has an essential singularity at ∞.
The function
∞
1 ∞
zn
e1/z + ez = + 1 +
n=1
n! z n n=0
n!
has essential singularities at z = 0 as well as z = ∞. The Laurent series on the right-
hand side of the equation above is valid in the annular region 0 < |z| < ∞.
The residue at infinity of a function f (z), however, is a little different from the
usual idea of a residue at a singularity at any point in the finite part of the complex
plane. I’ll return to this question shortly.
Accumulation point of poles: We have seen that the function cosec z has simple
poles at all integer multiples of π, i.e., at the set of points {z = nπ, n = 0, ±1, ±2, . . .}.
All these points lie on the x-axis. What is the nature of the singularity, if any, of cosec z
at z = ∞? The answer is provided by considering the Riemann sphere. Recall that
the mapping from the complex plane to the Riemann sphere is given by
z + z∗ z − z∗ |z|2 − 1
ξ1 = , ξ2 = , ξ3 = .
|z|2 + 1 i(|z|2 + 1) |z|2 + 1
The positive and negative segments of the x-axis on the complex plane map onto the
lines of longitude 0◦ and 180◦, respectively, on the Riemann sphere. Hence the poles
of cosec z lie on these lines of longitude, at the coordinates
2nπ n2 π 2 − 1
ξ1 = , ξ2 = 0 , ξ3 = , where n ∈ Z.
n π2 + 1
2 n2 π 2 + 1
Except for the south pole (corresponding to n = 0), all the poles lie in the northern
hemisphere (since π is larger than unity). It is evident that they get more and more
crowded together as n → ±∞, and approach the point (0 , 0 , 1), i.e., the north pole.
They become dense on the two longitude lines in an infinitesimal neighborhood of the
north pole. On the extended complex plane, z = ∞ is therefore an accumulation
point of poles. The singularity is no longer an isolated singularity.
Meromorphic function: An analytic function whose only singularities (if any) in the
finite part of the complex plane are isolated poles is called a meromorphic function.
Entire functions are trivially also meromorphic (as they have no singularities at all in
the finite part of the complex plane). Any rational function of z, i.e., a ratio of two poly-
nomials in z, is a meromorphic function. So are the functions cosec z, sec z, cosech z
and sech z. The function cosec (1/z) is not meromorphic, because it has an accumula-
tion point of poles at z = 0.
17
2. Identify the regions of the extended complex plane in which the functions listed be-
low are holomorphic. Also find the location, order and residue of each pole or essential
singularity of the functions, including the nature of the singularity (if any) at z = ∞.
2
(a) e−z (b) (1 − z 100 )/(1 − z) (c) (2z − 1)/z
(d) coth z − z −1 (f) (1 − cos z)/z 2
z
(e) (e)e
(g) 1/(ez + 1) (h) z cosec z (i) z cosec2 z
(j) (sin z − z cos z)/z 4 (k) 1/(2z − 1) (l) cosec (z 2 )
(m) 1/(sin z − cos z) (n) 1/(cos z − cosh z) (o) 1/(3z − 2z )
2
(p) 1/(ez − 1) (q) 1/(e2z − ez ) (r) exp [1/(z 2 − 1)]
∞
(s) cos (1/z) (t) (ez − 1)−2 (u) e−nz
n=0
∞
∞
∞
(v) z n /nn (w) z n /(n!)2 (x) [(1 − z)/(1 + z)]n
n=1 n=0 n=0
3. Obtain the Laurent series expansions of the following functions, valid in the regions
indicated. In each case, the expansion is understood to be in powers of (z − a), where
a is the center of the annular region concerned.
(a) (z + z −1 )1000 (0 < |z| < ∞) (b) z 2 e1/z (0 < |z| < ∞)
−1/z 2
(c) z −1 e (0 < |z| < ∞) (d) (z 2 + z −2 )100 (0 < |z| < ∞)
(e) e1/z (1 − z)−1 (0 < |z| < 1) (f) z 2 e1/(z−2) (0 < |z − 2| < ∞)
(g) 21/(z−1) z −2 (0 < |z − 1| < 1) (h) (z − 1)−1 (z − 2)−2 (0 < |z − 1| < 1).
Cauchy’s integral formula : Consider the contour integral C z n dz where n =
0, 1, . . . and C is a simple closed contour that encircles the origin once in the posi-
tive sense. As z n is an entire function, C may be distorted to a circle of arbitrary
radius r centered at the origin, and encircling it once in the positive sense. Since
dz = r eiθ i dθ on this circle, we have
2π
n n
z dz = i r dθ e(n+1)iθ = 0.
C 0
The vanishing of the contour integral is in any case immediately evident, even without
writing it out in the form above: the integrand z n is holomorphic everywhere inside
and on the contour C, so that C can be distorted till it shrinks to a point (i.e., you
18
can let r → 0), and the integral vanishes.
Now consider the contour integral C dz/z n+1 where n = 0, 1, . . . and C is again
a simple closed contour that encircles the origin once in the positive sense. As the
integrand z −n−1 has a pole of order (n + 1) at the origin, and no other singularities, C
can again be deformed to a circle of some radius r centered at the origin. Then
2π
dz i
n+1
= n dθ e−niθ .
C z r 0
As before, this integral vanishes identically an all cases except one: namely, when n = 0.
(Alternatively, we can now expand the circle outwards, and pass to the limit r → ∞.)
Only when n = 0, and in that case alone, do we get the nonzero value C dz/z = 2πi.
−1
Hence we have, for all integers n ∈ Z, (2πi) C
dz/z n+1
= δn , 0 . A little more
generally, if a is any complex number and C is a closed contour encircling a once in
the positive sense, then
1 dz
= δn , 0 .
2πi C (z − a)n+1
This is a most useful representation of the Kronecker delta. Cauchy’s Integral Theorem
and the whole of the so-called calculus of residues is based upon it.
Cauchy’s integral formula follows from the result above. Omitting the details,
the theorem states : Let C be a closed contour lying entirely in a region in which the
function f (z) is analytic, except for isolated poles and essential singularities at the
points zk , k = 1, 2, . . . . No singularities must lie on C itself. Then, if C encircles the
singularities once in the positive sense,
dz f (z) = (2πi) Res f (z).
C {poles zk }
The sign of the right-hand side is reversed if C encircles the singularities in the negative
or clockwise sense. If a singularity is encircled r times, the contribution to the integral
is ±r times the residue at that singularity, the negative sign being applicable when the
contour winds around the singularity in the negative sense. (r is called the winding
number of the contour at the singularity.)
19
Consider the integral from −R to R on the x-axis, and attach a semicircle of radius R
to close the contour either in the upper or lower half-plane. You now have the integral
of the function [(z 2 + a2 )(z 2 + b2 )]−1 over a closed contour C. By Cauchy’s theorem,
this integral is (apart from a factor ±2πi) the sum of the residues of the function at the
poles enclosed by C. Let R → ∞. The contribution from the semi-circle vanishes as
R → ∞, while the integral along the x-axis becomes the integral you want to evaluate.
∞
dx x2 π
(b) 2 2 3
= .
0 (x + a ) 16a3
Write the integral as 12 times the integral from −∞ to ∞ along the x-axis, and proceed
as in the preceding case. Note that the pole enclosed by C is a multiple pole in this case.
∞
dx x sin x
(c) 2 2
= π e−a .
−∞ (x + a )
In this case, the contour can’t be closed in either the upper or the lower half-plane
because sin z is a linear combination of eiz as well as e−iz , and one of the exponentials
will diverge as |z| → ∞ in each of the two half-planes.
∞ Therefore first write sin x as the
imaginary part of e , and consider the integral −∞ dx x e /(x + a2 ). It is clear that
ix ix 2
you can now close the contour in the upper half-plane. After evaluating the integral,
take its imaginary part. Alternatively, retain sin x in the integrand, and note that you
have to close the contour in different half-planes for the two exponential factors. Do
the integral both ways to check that the final results match.
∞
dx (x2 − x + 2) 5π
(d) 4 2
= .
−∞ x + 10x + 9 12
∞
dx sin x π
(e) = .
0 x 2
Extend the integral to one from −∞ to ∞, and consider the integrand eiz /z over a
contour that is closed in the upper half-plane. The pole at the origin must be avoided
by a small semi-circular indentation around z = 0. The rest of the integral along the
real axis is now a Cauchy principal value integral. After evaluation of the integral,
equating imaginary parts yields the result sought. Note that the principal value re-
striction can be dropped at this stage, because (sin x)/x does not have a singularity
at x = 0.
2π
dθ 2π
(f) =√ (a > b) .
0 a − b cos θ a2 − b2
20
Let z = eiθ . Then cos θ = 12 (z + z −1 ). The integration over θ becomes an integral over
z around the unit circle traversed once in the positive sense. Pick up the residue at
the pole of the integrand that lies inside the unit circle.
2π
2π
(g) dθ ecos θ cos (nθ − sin θ) = (n = 0, 1, . . .) .
0 n!
2π
The integral required is the real part of 0 dθ ecos θ ei(nθ−sin θ) . Set eiθ = z to convert
the integral to one around the unit circle in the z-plane. Pick up the residue of the
integrand at the singularity enclosed by the contour.
where the sum runs over the singularities at the points aj in the finite part of the
complex plane. Two noteworthy points follow at once.
(i) f (z) may have a nonzero residue at z = ∞ even if it is not singular at the point
at infinity!
For instance, the function (z − 1)−1 + (z − 2)−1 has simple poles at z = 1 (residue = 1)
and z = 2 (residue = 1), and is regular at z = ∞. And yet it has a nonzero residue
at infinity that is equal to the negative of the sum of its residues at z = 1 and z = 2,
namely, −2.
(ii) On the other hand, an entire function must have zero residue at infinity, even
though it is, in general, singular at that point!
Thus, a polynomial of order n, which has a pole of order n at z = ∞, must have
zero residue at that point. Likewise, an exponential such as ez , which has an essential
singularity at ∞, must have zero residue at that point.
21
circle c around the origin in the w-plane, with an infinitesimal radius 1/R, traversed
in the positive sense. Remembering that dz = −dw/w 2, this yields the formula
1 dw
Res f (z) = − f (1/w).
z=∞ 2πi c w 2
It follows that the residue at infnity of f (z) is the coefficient of w −1 in the Laurent
expansion of −(1/w 2)f (1/w) in powers of w. It is now easily checked that the residue
at infinity of (z − 1)−1 + (z − 2)−1 is indeed equal to −2, as found earlier. Similarly, it is
easy to verify that there is no term proportional to w −1 in the power series expansion
of −(1/w 2 )f (1/w) when f (z) is any polynomial or exponential like ez .
Note that this result does not depend on the details of the locations or the multiplici-
ties of the zeroes of qn+1 (z).
6. Summation of series using contour integration: The fact that π cot πz has a
simple pole at each integer z = n, with residue equal to +1, can be used to sum certain
infinite series. Similarly, the fact that π cosec πz has simple poles at all integers z = n,
with residue equal to (−1)n , can be used to sum certain series in which the signs of
the terms alternate. The method works when the summand is an even function of n,
so that the series can be written as a sum over both positive and negative integers.
Evaluate the following infinite sums by the contour integration method. (The values
of the sums are also given, so that you can check your answers.)
∞
1 π 1
(a) S(a) = 2 2
= coth πa − (a > 0).
n=1
n +a 2a πa
First write the sum required as 12 times the sum over positive as well as negative integers
n. Each terms in the sum is just the residue of the function f (z) = π cot (πz)/(z 2 + a2 )
at its simple pole at z = n. Hence it is (2πi)−1 times the contour integral of f (z) over a
small circle encircling z = n in the positive sense. Merge all these little circles together,
to get two ‘hairpin’ contours straddling the x-axis from −∞ to −1 and from 1 to ∞,
respectively. These can be made part of a single closed contour by attaching large
semi-circles in the upper and lower half-planes. This contour now encircles just three
22
poles of f (z), namely, those at z = 0 and z = ±ia, but in the clockwise or negative
sense. Pick up the residues at these poles and simplify the resulting expression to
obtain the final result given above.
∞
1 π2
(b) = .
n=1
n2 6
This sum is equal to ζ(2), where ζ(s) = ∞ s
n=1 1/n is the Riemann zeta function.
The same method as that used in part (a) above will lead to the result quoted. Note
that there is a pole of order 3 at z = 0 in this instance.
Verify the result above by passing to the limit a → 0 in the sum S(a) found in part
(a). You will need to use the fact that cot z = z −1 + 13 z + O(z 3 ) in the neighborhood
of z = 0.
1
(c) Use the fact that ζ(2) = 6
π 2 to show (by elementary means) that
∞
(−1)n−1 π2
∞
1 π2
= and = .
n=1
n2 12 n=1
(2n − 1)2 8
∞
(−1)n−1 π 1 1
(d) 2 2
= − .
n=1
n +a 2a πa sinh πa
Repeat the procedure used in part (a), but with cosec (πz) instead of cot (πz) in the
definition of the integrand f (z). Once again, verify the result you obtain by noting
that
∞
(−1)n−1 ∞
1
2 2
= S(a) − 2 S( 2 a) where S(a) =
1 1
.
n=1
n +a n=1
n + a2
2
Finally, pass to the limit a → 0 to check that ∞ n=1 (−1)
n−1 1 2
/n2 = 12 π .
∞
1 π πa 2
(d) 2 2 2
= 3 coth πa + 2 − .
n=1
(n + a ) 4a sinh πa πa
Use the same procedure as in part (a) to establish this result. Verify the answer directly
from S(a) by observing that
∞
1 1 dS(a)
=− .
n=1
(n2 2
+a ) 2 2a da
23
(e) Pass to the limit a → 0 in the foregoing result to establish that
∞
1 π4
∞
(−1)n−1 7π 4
ζ(4) = 4
= , and hence = .
n=1
n 90 n=1
n4 720
∞
1 π 1
(f) = (b coth πa − a coth πb) − .
n=1
(n2 + a2 )(n2 + b2 ) 2ab(b2 − a2 ) 2a2 b2
Verify the result by noting that the sum required is [S(a) − S(b)]/(b2 − a2 ).
∞
1
(g) Find the sum (α > 0).
n=1
n4 + α4
It is evident that we must now consider the contour integral of the function π cot (πz)/(z 4 +
α4 ). Check your answer by writing (n4 + α4 ) as (n2 + a2 )(n2 + b2 ), where a = α eiπ/4
and b = α e3πi/4 = −α e−iπ/4 , and using (the analytic continuation of) the result of
part (f).
The set {cn } in this case is a Fibonacci sequence of numbers. (Each member of the
sequence is the sum of the preceding two numbers.) The answer for cn is, in this case,
1 √ √
cn = √ ( 5 + 1)n − (−1)n ( 5 − 1)n , n = 0, 1, . . . .
2n 5
24
√
As you know, the irrational number τ = 12 ( 5 + 1) is called the golden mean, and
has a large number√of interesting properties. So do the Fibonacci numbers, since
cn = [τ n − (−τ )−n ]/ 5. The result above shows that
• the asymptotic (large-n) growth of the Fibonacci sequence is exponential, ∼ eλn ,
where λ = ln τ 0.4812 .
(b) cn+2 − cn+1 − 2cn = 0 , n ≥ 0, c0 = 1 , c1 = 2.
25
The Mittag-Leffler expansion of f (z) is then of the form
f (z) = sj (z) + g(z),
j
where g(z) is an entire function. When the number of poles is finite, the sum over
j presents no problem. This is the case when f (z) is a rational function, i.e., the
ratio of two polynomials. In such cases the Mittag-Leffler expansion is nothing but
the resolution of the function into partial fractions. For instance, the Mittag-Leffler
expansion of the function f (z) = (z − a)/(z − b)2 is just
z−a b−a 1
= + .
(z − b) 2 (z − b) 2 z−b
When the summation over j involves an infinite number of terms, however, the ques-
tion of the convergence of the sum arises.
Although the infinite series was originally summed for real positive values of a, we may
regard the equality above as a relation between two analytic functions of the complex
variable a, by virtue of the principle of analytic continuation. Set a = iz and note
that coth (iπz) = −i cot (πz). The relation above can be written as
1 ∞
z ∞
z
π cot (πz) = + 2 = .
z n=1
z −n
2 2
n=−∞
z − n2
2
We’re almost there, because this expansion seems to express the meromorphic function
π cot (πz) as a sum over the singular parts at its poles (which are located at all integer
values of z). But the poles of cot (πz) are simple poles, whereas the summand in the
last equation involves a quadratic function of z in the denominator. It is a simple
matter to write ∞
∞
2z 1 1
2 − n2
= + .
n=1
z n=1
z − n z + n
But we cannot go on to split the right-hand side into two sums and write it as
∞
1
∞
1
∞
1
−1
1
+ = + ,
n=1
z − n n=1 z + n n=1 z − n n=−∞ z − n
26
because each of the two individual infinite series in the last line above diverges! To
find the correct Mittag-Leffler expansion, let’s go back to
∞
z 1
∞
z
π cot (πz) = = + .
n=−∞
z −n
2 2 z n=−∞ z − n2
2
n=0
But
∞
1
∞ ∞
z 1 2z 1 1
= = +
n=−∞
z − n2
2 2
n=−∞
z − n2
2 2
n=−∞
z−n z+n
n=0 n=0 n=0
∞
1 1 1 1
= 1
+ + −
2
n=−∞
z−n n z+n n
n=0
∞
1 1
= + ,
n=−∞
z−n n
n=0
where we have used (in writing the last line) the fact that the sum over all nonzero
integers remains unchanged if n → −n. Hence we obtain
∞
1 1 1
π cot (πz) = + + .
z n=−∞ z − n n
n=0
This is the Mittag-Leffler expansion of the meromorphic function π cot (πz). Note
that the term 1/n inside the large brackets in the summand on the right-hand side
is necessary. Without it, the sum over n of 1/(z − n) would diverge. With the 1/n
term present, the summand is sufficiently well-behaved5 to guarantee convergence of
the series.
Moreover, the series can be differentiated term by term with respect to z. The
result is the Mittag-Leffler expansion of the function cosec2 (πz), which reads
π2 ∞
1
= .
sin πz n=−∞ (z − n)2
2
5
It falls off like 1/n2 as |n| → ∞.
27
3 Linear response; dispersion relations
Causal, linear, retarded response: Under very general conditions, the physical re-
sponse of a system to an applied time-dependent stimulus satisfies certain basic criteria.
Let us denote the stimulus (or ‘force’) by F (t), and the response by R(t). Examples of
such stimulus-response pairs are : mechanical force and displacement; electric field and
polarization; magnetic field and magnetization; stress and strain; electromotive force
and current; and so on. For notational simplicity, I have suppressed the appropriate
indices in the stimulus and the response when these are vectors, tensors, etc. in what
follows.
Under fairly general conditions, the response at any instant of time t may be as-
sumed to meet three basic requirements:
(i) Causality: The response at time t depends on the force history at all earlier
instants of time, but not later ones. This is the principle of causality, which says that
the effect cannot precede the cause.
(ii) Linearity: The response is assumed to be linear in the applied force. This implies
at once that the superposition principle holds good.
(iii) Retarded response: The effect of a force F (t ) applied at the instant t on the
response R(t) at a later instant t depends only on the elapsed time interval (t − t ),
rather than on both t and t individually. This means that the location of the actual
origin of the time coordinate, the particular instant at which we set t = 0, does not
matter.
Taking the force to be applied from t = −∞ onward (this will automatically incorporate
all other starting times), the most general linear functional of F (t) subject to these
requirements is given by
t
R(t) = dt φ(t − t ) F (t ).
−∞
Note that the upper limit of integration over t is t, in accordance with the require-
ment of causality. The quantity φ(t − t ) is called the response function. The fact
that it is a function of the time difference (t − t ) indicates that we are dealing with a
retarded response. The quantity φ(t) represents the ‘weight’ with which a force applied
at t = 0 contributes to the response R(t) at any later time t. We may expect it to be
a decreasing (or at least a non-increasing) function of its argument, such as a decaying
exponential. But it could oscillate as a function of t, although we would expect it to
do so with decreasing amplitude, in most instances. These are not strict requirements,
of course, but they seem to be plausible on physical grounds.
28
Dynamic susceptibility: It is natural to decompose general time-dependent func-
tions such as F (t) and R(t) into their Fourier (or frequency) components. We have
∞ ∞
−iωt
F (t) = dω e F (ω) and R(t) = dω e−iωt R(ω).
−∞ −∞
1. Use these relations in the foregoing relation between R and F to show that
∞
R(ω) = χ(ω) F(ω), where χ(ω) = dt eiωt φ(t).
0
where ∗ denotes the complex conjugate. I have added the phrase ‘for all real ω’ has
been added because it makes sense, as you’ll see shortly, to regard of χ(ω) as an
analytic function of the complex variable ω, although the physically accessible values
of the frequency are of course real and non-negative. It follows at once that, for all
real ω,
Re χ(−ω) = Re χ(ω) and Im χ(−ω) = −Im χ(ω).
The real part of the dynamic susceptibility is an even function of the frequency, while
the imaginary part is an odd function of the frequency. You can, of course, write down
29
these properties directly by setting eiωt = (cos ωt + i sin ωt) in the definition of χ(ω).
Consider the formula for χ(ω) as the one-sided Fourier transform of φ(t). We have
assumed that the integral exists for physical (i.e., real, non-negative) values of the
frequency ω. It is obvious that it will continue to do so even if ω is complex, pro-
vided Im ω is positive. This is because the factor eiωt then leads to an extra damping
factor e−(Im ω) t . Such a factor can only improve the convergence of the integral, since
the integration over t is restricted to non-negative values. We may therefore conclude
that the dynamic susceptibility can be analytically continued to the upper half-plane6
(UHP for short) in ω. The defining expression for χ(ω) provides a representation for
this analytic function for all Im ω ≥ 0.
Given that the dynamic susceptibility is analytic in the UHP, we can extend the
relation χ(−ω) = χ∗ (ω) (that holds good for real ω) to complex values of ω. We have
χ(−ω ∗ ) = χ∗ (ω), Im ω ≥ 0.
Note that if ω lies in the UHP, so does −ω ∗ . Hence the arguments of the functions on
both sides of the equation above do lie in the region in which we are guaranteed that
the dynamic susceptibility is analytic. On general grounds, and without further input,
we really cannot say very much about its possible behavior in the lower half of the
complex ω-plane. But we know that an analytic function of a complex variable cannot
be holomorphic at all points of the extended complex plane unless it is just a constant,
as we know from Liouville’s Theorem. In general, therefore, the dynamic susceptibility
will have singularities in the lower half-plane in ω.
Dispersion relations : The analyticity of χ(ω) in the UHP enables us to derive cer-
tain relations between its real and imaginary parts. Let us assume that |χ(ω)| → 0
as ω → ∞ along any direction in the UHP. This is a physically plausible assumption
to make in most circumstances, for the following reason. Even for real ω, we expect
the susceptibility to vanish as the frequency becomes very large, because the inertia
present in any system will not permit it to respond to a sinusoidal applied force os-
cillating at a frequency much higher than all the natural frequencies present in the
system. And when ω is a complex number with a positive imaginary part, the factor
eiωt in the formula for the susceptibility lends an extra damping factor e− (Im ω) t . The
assumption is thus a reasonable one to make. However, it is not absolutely essential,
6
Causality ensures that the Fourier-Laplace transform of a causal response function is analytic in a
half-plane in ω. Whether this is the upper half-plane (UHP) or lower half-plane (LHP) depends on the
Fourier transform convention chosen. With our particular convention in which a function f (t) of the
∞
time is expanded as −∞ dω f(ω) e−iωt , it is the upper half of the ω-plane in which the susceptibility
is analytic. Had we chosen the opposite sign convention and used the factor e+iωt in the expansion
above, the region of analyticity would have been the lower half-plane in ω. The convention I have
adopted is the more commonly used one, at least in physics.
30
and can be relaxed, as you’ll see further on.
χ(ω )
f (ω ) =
ω −ω
as a function of the complex variable ω . This function is analytic everywhere in the
upper half-plane in ω , as well on the real axis in that variable, except for a simple
pole at ω = ω located on the real axis. By Cauchy’s integral theorem its integral
over any closed contour C lying entirely in the upper half-plane is identically equal to
zero Without changing the value of the integral (namely, zero), the contour C can be
expanded to the contour C that comprises the following:
(i) a large semicircle of radius R in the UHP,
(ii) a line integral on the real axis running from −R to ω − ,
(iii) a small semicircle, from ω − to ω + , lying in the UHP so as to avoid the simple
pole of the integrand, and, finally,
(iv) a line integral from ω + to R.
Thus
χ(ω )
dω f (ω ) = dω = 0.
C C ω−ω
In the limit R → ∞, the contribution from the large semicircle vanishes, because f (ω )
vanishes faster than 1/ω as ω → ∞ along all directions in the UHP: this is ensured by
the fact that χ(ω ) → 0 as |ω | → ∞. On the small semicircle, we have ω = ω + eiθ ,
where θ runs from π to 0. Therefore, in the limit → 0, the contribution from the
small semicircle tends to −iπχ(ω). Hence we get
ω−
∞
χ(ω ) χ(ω )
lim dω + dω − iπχ(ω) = 0.
→0 −∞ ω−ω ω+
ω−ω
But the limit on the left-hand side of this equation is just the Cauchy principal value
integral of the integrand χ(ω )/(ω − ω). Recall that the Cauchy principal value is a
specific prescription for avoiding the divergence or infinity that would otherwise arise,
owing to the singularity (a simple
pole) of the integrand at ω = ω. Denoting this
principal value integral by P (· · · ), we get
∞
i χ(ω )
χ(ω) = − P dω .
π −∞ ω −ω
Note that this equation expresses the susceptibility at a real frequency as a certain
weighted sum of the susceptibility over all other real frequencies. No complex frequen-
cies appear anywhere in this formula. We made an excursion into the upper half of
31
the complex ω-plane. This was made possible by the analyticity properties of the sus-
ceptibility. But we have returned to the real axis, bringing back the last equation with
us. Equating the respective real and imaginary parts of the two sides of this equation,
we get
∞ ∞
1 Im χ(ω ) 1 Re χ(ω )
Re χ(ω) = P dω and Im χ(ω) = − P dω .
π −∞ ω−ω π −∞ ω −ω
These formulas are called dispersion relations (the origin of the term is given below).
They imply that the two real functions of a real argument, Re χ(ω) and Im χ(ω) , form
a Hilbert transform pair.
2. The dispersions relations above still involve integrals over negative as well as pos-
itive frequencies, whereas physically accessible frequencies are non-negative. But the
symmetry properties of Re χ(ω) and Im χ(ω) can be used to restrict the range of
integration to physically accessible frequencies, namely, 0 ≤ ω < ∞. Show that
∞ ∞
2 ω Im χ(ω ) 2ω Re χ(ω )
Re χ(ω) = P dω and Im χ(ω) = − P dω .
π 0 ω 2 − ω2 π 0 ω 2 − ω2
The term ‘dispersion relation’ originates from the fact that they were first derived
in the context of optics : the real and imaginary parts of the frequency-dependent re-
fractive index of an optical medium are, respectively, measures of the dispersion and
absorption of radiation by the medium. These are not independent quantities, but are
related to each other via dispersion relations, which are also called Kramers-Kronig
relations in physics. As I have already mentioned, they are a direct consequence of
the principle of causality as applied to response functions.
Admittance of an LCR circuit : Consider a series LCR circuit. If the applied emf
is V (t), the current I(t) is given by the equation
dI 1 t
L + RI + I(t ) dt = V (t).
dt C −∞
Now use the Fourier expansions
∞ ∞
−iωt
V (t) = dω e V (ω) and I(t) = dω e−iωt I(ω).
−∞ −∞
7
A function g(z) is a real analytic function of z = x + iy if it is (i) an analytic function of z,
and (ii) real when its argument is real, i.e., g(x) is real.
32
Inserting these in the equation for I(t), we find that the last term on the left-hand side
becomes (after a change of integration variable from t to τ = t − t )
∞
1 ∞ −iωt
dω e I(ω) dτ eiωτ .
C −∞ 0
The problem is that the last integral above diverges, i.e., it is formally infinite. We
should really have used Laplace transforms instead of Fourier ∞transforms to tackle the
iωτ
problem under ∞consideration! Had we done so, the integral 0 dτ e would have been
−sτ
replaced by 0 dτ e , with s in a region such that Re s large enough to guarantee
the convergence of the integral involved. In this case this is simply Re s > 0, and
the value of the integral
∞ is just 1/s. Let’s suppose that this has been done, so that
we may interpret 0 dτ eiωτ as standing for the analytic continuation to s = −iω of
the corresponding Laplace transform. This gives the value 1/(−iω) = i/ω for the
apparently divergent integral. With this small technicality taken care of, the Fourier
transforms of the current and the voltage are related to each other according to
i
−iωL + R + I(ω) = V (ω).
ωC
The quantity in the brackets on the left-hand side is of course the complex impedance
of the circuit, usually denoted by Z(ω). The corresponding dynamic susceptibility
is nothing but the reciprocal of the impedance, namely, the complex admittance,
customarily denoted by Y (ω). We have
I(ω) = Y (ω) V (ω),
where −1
i iω
Y (ω) = −iωL + R + = .
ωC L (ω + iγ ω − ω02)
2
√
You will recognize γ −1 = L/R as the time constant of an LR circuit, and ω0 = 1/ LC
as the natural frequency of an undamped LC circuit. Observe that Y (ω) is analytic in
the upper half-plane in ω, as required. Moreover, in this simple case we know its form
explicitly. It has two simple poles in the lower half-plane, at
ω± = − 2 iγ ± ω02 − 14 γ 2
1
respectively. Further, Y (ω) vanishes like ω −2 as |ω| → ∞ in the upper half-plane. Its
real and imaginary parts must therefore satisfy (unsubtracted) dispersion relations.
γω 2 iω(ω 2 − ω02 )
Re Y (ω) = and Im Y (ω) =
(ω 2 − ω02 )2 + ω 2 L2 (ω 2 − ω02)2 + ω 2 L2
33
satisfy the dispersions relations
∞ ∞
2 ω Im Y (ω ) 2ω Re Y (ω )
Re Y (ω) = P dω , Im Y (ω) = − P dω 2 .
π 0 ω 2 − ω2 π 0 ω − ω2
To get over the problem, we assume that the value of χ(ω) at some particular real
value of ω0 of the frequency is known. Then, instead of the function χ(ω )/(ω − ω),
consider
χ(ω ) − χ(ω0 )
f (ω ) = .
(ω − ω0 )(ω − ω)
It is evident that this function does vanish faster than 1/ω as |ω | → ∞, owing to
the extra factor (ω − ω0 )−1 . Moreover, it is analytic everywhere in the upper half
ω -plane and on the real axis, except for a simple pole at ω = ω, as before. Observe
that it does not have any singularity at ω = ω0 . This is the reason for subtracting
χ(ω0 ) from χ(ω ) in the numerator. Repeat the earlier derivation to show that
∞
i χ(ω ) − χ(ω0 )
χ(ω) = χ(ω0 ) − (ω − ω0 ) P dω .
π −∞ (ω − ω0 )(ω − ω)
Hence obtain the dispersion relations satisfied by the real and imaginary parts of χ(ω)
in this case. These are called once-subtracted dispersion relations, and ω0 is the
‘point of subtraction’. From the mathematical point of view, it should be evident
that the procedure given above can be extended to cover situations when the analytic
function χ(ω) actually diverges as ω → ∞ in the upper half-plane. For instance, if
χ(ω) ∼ ω n−1 asymptotically, where n is a positive integer, we can write down an n-fold
subtracted dispersion relation for it. The latter will require n constants as inputs.
These could be, for instance, the values of χ(ω) at n different frequencies or points of
subtraction.
Hilbert transform pairs : Given a real function g(x) of the real variable x, its
Hilbert transform is defined as
∞
1 g(x )
h(x) = P dx .
π −∞ x −x
34
The inverse transform is given by
∞
1 h(x )
g(x) = − P dx .
π −∞ x −x
Eliminating h, we have
∞ ∞
1 g(x )
g(x) = − 2 P dx P dx
π −∞ −∞ (x − x)(x − x )
Hence the quantity in the curly brackets in the equation above must be the unit
operator in function space, i.e., we must have
∞
1 dx
P = δ(x − x ).
−∞ (x − x)(x − x )
π 2
6. Discrete relaxation spectrum : As we have seen, the real and imaginary parts
of the dynamic susceptibility constitute a Hilbert transform pair. Given that
n
σj
Re χ(ω) =
j=1
1 + ω 2 τj2
where σj and τj are positive constants, use the dispersion relation to show that
n
ω σj τj
Im χ(ω) = 2 τ2
.
j=1
1 + ω j
Hence
n
σj
χ(ω) = .
j=1
1 − iω τj
35
Thus χ(ω) has a set of poles in the lower half-plane in ω, at ω = −i/τj , 1 ≤ j ≤ n.
Physically, this corresponds to a discrete relaxation spectrum comprising the re-
laxation times τj , contributing with corresponding weights σj . The response function
φ(t) itself is a weighted sum of decaying exponentials e−t/τj . In a physical situation in
which there is essentially just a single relaxation time τ (or when it is a good approxi-
mation to assume that this is so), we have the well-known case of Debye relaxation.
In this instance, the so-called Cole-Cole plot or the Argand diagram of χ(ω), i.e., a
plot of (the suitably scaled variables) Im χ versus Re χ, obtained by eliminating ω, is
the upper half of semi-circle of radius 12 centered at ( 12 , 0). Departures of the Cole-Cole
plot from this curve serve as an indication that the relaxation is not of the single-
exponential form.
36
4 Analytic continuation and the gamma function
The gamma function : In practice, there are numerous methods of analytically
continuing a function from a region in which a local representation of the function is
available, to a larger region. Some of these are: chains of rearranged power series; the
Schwarz reflection principle (for real analytic functions); the use of the ‘permanence’
of functional equations; summability methods for power series; integral representations
for functions in terms of contour integrals, using the distortability of the contours; and
so on.
A good illustration is provided by the gamma function, which may be defined (to
start with) by the integral
∞
Γ(z) = dt tz−1 e−t , Re z > 0.
0
As you know, Γ(n + 1) = n! for positive integer values of n. The integral above serves
also to define 0! as unity. The question is, how do we analytically continue the gamma
function to the region Re z ≤ 0 in the complex z-plane?
It is obvious that the problem with the convergence of the integral above for Re z ≤
0 arises from the lower limit of integration in t. The behavior of the integrand at this
end-point can be improved by integration by parts. Keeping Re z > 0, integrate by
parts to obtain
∞
Γ(z) = dt tz−1 e−t (Re z > 0)
0
z −t t=∞
t e 1 ∞ z −t
= + dt t e (Re z > 0)
z z 0
t=0
=0 because Re z>0
But the final integral on the right-hand side converges in the extended region Re z >
−1, while the factor 1/z multiplying it suggests that the gamma
∞ function itself has a
z −t
simple pole at z = 0. And indeed it does, because the integral 0 dt t e is finite and
nonzero at z = 0. It is equal to 1 when z = 0, so that we may conclude that Γ(z) has
a simple pole at z = 0 with residue equal to unity. We have thus extended the region
in which we have an explicit representation of the gamma function by the extra strip
−1 < Re z ≤ 0, and we may write
1 ∞ z −t
Γ(z) = dt t e , Re z > −1.
z 0
Continuing this procedure of integrating by parts, it is easy to see that we have
∞
1
Γ(z) = dt tz+n e−t , Re z > −n − 1.
z(z + 1) · · · (z + n) 0
37
The integral on the right-hand side converges for Re z > −n−1, for any n = 0, 1, 2, . . . ,
while the factor multiplying it makes the poles at z = 0, 1, . . . , −n explicit. We have
thus achieved an analytic continuation of the gamma function, strip by strip, to the
region Re z > −n − 1, starting with a representation for the function that was only
valid in the region Re z > 0. In principle, this can be done for arbitrarily large value
of n. The underlying mechanism is summarized in the functional equation satisfied by
the gamma function, namely,
1
Γ(z + 1) = z Γ(z) or Γ(z) = Γ(z + 1).
z
The ‘permanence’ of this equation (i.e., its validity for all z) can be used to analytically
continue the gamma function arbitrarily far to the left of the imaginary axis in the
complex z-plane.
1. Connection with the Gaussian integral: The gamma function helps us express
a number of very useful integrals in closed form.
(a) Show, by a simple change of variables, that
∞
2
du un e−au = 12 Γ 12 (n + 1) a−(n+1)/2 , a > 0, n > −1.
0
Note, in particular, that n need not be an integer in this formula. In fact, as you can
readily guess, the formula continues to be valid for complex values of n and a, provided
Re n > −1 and Re a > 0. Setting n = 0 and a = 1 in the formula, we may turn it
around to obtain ∞
1 2 √
Γ 2 =2 du e−u = π,
0
where we have made use of the known result for the basic Gaussian integral.
(b) It follows at once that the value of the gamma function of a half-odd-integer can
be written down explicitly. Use the functional equation for the gamma function
to show that
√ √
(2n)! π 2n
n 2 n! π
Γ n+ 2 =1
2n
and Γ −n + 2 = (−1)
1
, n = 0, 1, 2, . . .
2 n! (2n)!
Analytic properties of the gamma function: It is clear from the analytic con-
tinued form of the gamma function (obtained either by integration by parts, or from
the functional equation) that Γ(z) has simple poles at all the non-positive integers,
n = 0, −1, −2, . . . . Its residue at z = −n is equal to (−1)n /n! . Other than these
38
simple poles, it has no other singularities in the finite part of the complex plane. It
is therefore a meromorphic function. It is then natural to ask for the Mittag-Leffler
expansion of Γ(z)—namely, a representation that explicitly separates out the contri-
bution from all the poles of the function from the part that is an entire function.
2. Mittag-Leffler expansion of Γ(z): Note that the pole at z = −n arises from the
behavior of the factor tz+n−1 at the lower limit of integration, t = 0, in the integral
representation of Γ(z). Write
∞ 1 ∞
z−1 −t z−1 −t
Γ(z) = dt t e = dt t e + dt tz−1 e−t .
0 0 1
The second integral on the right-hand side in the last equation above does not have
any singularities in the region |z| < ∞, and is therefore an entire function of z. In the
integral from t = 0 to t = 1, expand e−t in powers of t and integrate term by term, to
obtain the Mittag-Leffler expansion of the gamma function :
∞ ∞
(−1)n
Γ(z) = + dt tz−1 e−t .
n! (z + n)
1
n=0
entire function
sum over pole terms
d Γ (z)
ψ(z) = Γ(z) = .
dz Γ(z)
The functional relation Γ(z + 1) = z Γ(z) leads at once to the difference equation
1
ψ(z + 1) − ψ(z) = .
z
From the fact that
(−1)n
Γ(z) = + regular part
n! (z + n)
in the neighborhood of its simple pole at z = −n, it follows that Γ (z) has a double
pole at z = −n, and has the behavior
(−1)n
Γ (z) = − + regular part.
n! (z + n)2
Hence show that ψ(z) has simple poles at n = 0, −1, −2, . . . , with residue equal to
−1 at each pole.
39
Infinite product
representation for the gamma function: Recall that the har-
−1
monic series N n=1 n diverges logarithmically as N → ∞. That is, the sum has a
leading asymptotic behavior ∼ ln N for large
values of N. The question that arises
N −1
naturally is: does the difference n=1 n − ln N tend to a finite limit as N → ∞?
It turns out that
N
lim n−1 − ln N = γ,
N →∞
n=1
40
where m and n are positive integers. It is clear, however, that the same integral
representation can be used to define the beta function as a function of two complex
variables z and w, say, according to
1
B(z, w) = dt tz−1 (1 − t)w−1 , Re z > 0 and Re w > 0.
0
The actual structure of the beta function is clarified by relating it to the gamma func-
tion as follows.
which converges in the region Re z > 0 and Re w > 0. Using the formula connecting
the Gaussian integral to the gamma function,
Hence, equating the two expressions for I(z, w), arrive at the basic relationship
Γ(z) Γ(w)
B(z, w) = .
Γ(z + w)
As always, this equation must be regarded as one between two analytic functions, valid
for all values of the arguments z and w. It leads to a number of important properties
41
of the gamma function.
The integral on the right-hand side can be evaluated by contour integration.9 The
result is
π
Γ(z) Γ(1 − z) = .
sin πz
This is the formula sought. Once again, it is a relation between analytic functions.
Therefore, by analytic continuation, it is valid for all z. Note that cosec (πz) has a
simple pole at every integer value of z. On the left-hand side, Γ(z) supplies the poles
at zero and the negative integers, while Γ(1 − z) has poles at the positive integers.
Taking logarithmic derivatives on both sides, it follows that
5. Using the fact that Γ(z) is a real analytic function of z, i.e., it is real when z is real,
show that 1
Γ + iy 2 = π sech (πy).
2
But the integrand in the above is symmetric about the mid-point of integration, t = 12 .
Hence 1/2
B(z, z) = 2 dt [t (1 − t)]z−1 .
0
9
You’ll see how this can be done after we discuss multivalued functions and branch points.
42
Change variables to u = 4t(1 − t), to get
B(z, z) = 21−2z B z, 12 .
Now use the relation between the beta and gamma functions to obtain Legendre’s
duplication formula, namely,
22z−1
Γ(2z) = √ Γ(z) Γ z + 12 .
π
The formula is actually a special case of a multiplication theorem for the gamma
function, which reads
n−1 j
(1−n)/2 nz− 21
Γ(nz) = (2π) n Γ z+ , n = 1, 2, . . . .
j=0
k
Use this relation, and the difference equation for ψ(z), to establish the following:
n
1
(a) ψ(z + n) = ψ(z) + .
j=1
(z + j − 1)
n
1
(b) ψ(n + 1) = −γ + .
j=1
j
1
(c) ψ 2
= −γ − 2 ln 2 .
n
1
(d) ψ n + 1
= −γ − 2 ln 2 + 2 .
2
j=1
(2j − 1)
43
5 Möbius transformations
Conformal mapping: Every analytic function f (z) is a map of some region R of the
complex plane to some region R of the complex plane. In other words, given a complex
number z ∈ R, the map f : z → w produces another complex number w ∈ R . Under
such a mapping, points in R map to points in R , curves to curves, and areas to areas,
in general.
• The special feature of an analytic function f (z) is that this map is conformal.
That is, it preserves the angles between curves.
Conformal mapping is a standard and very well-studied topic in complex analysis. It
is also one that has a number of practical uses—for instance, in fluid dynamics and
aerodynamics. We will not go into conformal mapping per se here. Instead, in this
chapter, the focus is on one specific kind of conformal map, because it is a very special
one in a sense that will become clear shortly. It is also of fundamental importance in
many areas of mathematical physics.
f (z) = az + b,
where a and b are complex constants. Further, the point at ∞ is mapped to itself under
this map, i.e., ∞ is a fixed point of this map. In terms of the real and imaginary
parts of z = x + iy, this map amounts to the linear transformations
x → a1 x − a2 y + b1 , y → a2 x + a1 y + b2 ,
where a = a1 + ia2 and b = b1 + ib2 . In geometrical terms, this map is made up (in
general) of a rotation, dilation and shear in the xy-plane, followed by a shift of the
origin. Interesting as it is, the map is quite simple, and the matter appears to end
44
here. But this is not quite so, as you will see below.
The fixed points of a Möbius transformation, i.e., points that map to themselves
under the transformation, are of fundamental importance. They help classify the
10
The latter name is more commonly used when the transformation is regarded as a mapping of the
Riemann sphere, and I will use this term. There are several other names (!) for this transformation—
linear fractional transformation, homographic transformation, projective transformtion, and bilinear
transformation.
45
transformations into distinct types, as you’ll see later. The fixed points are the roots
of the equation f (z) = z, or
cz 2 + (d − a)z − b = 0.
It is immediately clear that a Möbius transformation can have at most two fixed points,
which I’ll denote by ξ1 and ξ2 . (The only exception is the identity transformation,
which of course maps every point to itself.) Two cases arise, each with a sub-case, so
that there are four possibilities:
(iv) Both fixed points at ∞: If c = 0 and further a = d, then the linear transfor-
mation is merely a shift of the origin,
b
w=z+ = z ± b,
a
since a2 = 1 in this case. The fixed points then coincide at infinity,
ξ1 = ξ2 = ∞.
46
The cross-ratio of four points is a basic concept in geometry. It plays a fundamen-
tal role in the subject of projective geometry. It is extremely useful in the context of
Möbius transformations, as many results can be established quite easily with the help
of its properties.
Let z1 , z2 , z3 and z4 be an ordered set of four distinct points in the complex plane.
The cross-ratio of these points is defined as
(Draw a figure showing the distances involved in the cross-ratio, for a general configu-
ration of points.) It is obvious that the cross-ratio of four points depends on the order
in which they are specified. Hence there are 4! = 24 possible cross-ratios that can be
associated with an unordered set of four points. However, there are several obvious
symmetries among these, and the 24 cross-ratios are related to each other. Let’s write
[zi , zj ; zk , zl ] as [ij; kl], for brevity. It is easy to see that
[ij; kl] = 1/[ij; lk] = 1/[ji; kl] = [ji; lk] = [kl; ij].
The identities
are also easily established. Using these, it is possible to express all 24 cross-ratios
among any set of 4 points in terms of any one cross-ratio (see below). If one of the four
points in a cross-ratio is the point at infinity, the cross-ratio is defined by a limiting
process. For instance, if z1 = ∞, we have
(z2 − z4 )
[∞ , z2 ; z3 , z4 ] = lim [z1 , z2 ; z3 , z4 ] = .
z1 →∞ (z2 − z3 )
• A Möbius transformation leaves the cross-ratio of any four arbitrary points un-
altered in value.
(a) Find the relations between the 24 cross-ratios associated with any four distinct
points (z1 , z2 , z3 , z4 ) ∈ C.
47
From the invariance of the cross-ratio under a Möbius transformation, the following
property can be established:
• Given two sets of three distinct points (z1 , z2 , z3 ) and (w1 , w2 , w3 ) on the Rie-
mann sphere, there exists a unique Möbius transformation that maps zi to wi for
i = 1, 2 and 3.
There are several ways to prove this assertion. We won’t go into the complete proof
here, but it is possible to construct the Möbius transformation concerned quite easily:
• In the complex plane, a Möbius transformation maps circles and straight lines
to circles and straight lines. (Recall that a straight line in the complex plane is
a circle passing through the point of projection on the Riemann sphere.)
3. Let z1 , z2 , z3 be three arbitrary finite points in the complex plane. Show that the
Möbius transformation that maps the circle passing through these three points (or
the straight line on which they lie, if they are collinear) to the real axis such that
z1 → 0, z2 → 1, z3 → ∞ is
(z − z1 )(z3 − z2 )
w= .
(z − z3 )(z1 − z2 )
Hint: Use the fact that we must have [z , z1 ; z2 , z3 ] = [w , 0 ; 1 , ∞] under the trans-
formation sought. Solve this equation to find w as a function of z.
48
(i) Finite, distinct fixed points ξ1 and ξ2 : Consider the four points z, ∞, ξ1 and
ξ2 , where z is variable. Under the Möbius transformation, these points are mapped
according to
a
z → w, ∞ → , ξ1 → ξ1 , ξ2 → ξ2 .
c
Since the cross-ratio does not change under the mapping, we have
Writing this out explicitly, we have the following result. Define the constant
def. a − c ξ1 a + d − (a + d)2 − 4
K = = .
a − c ξ2 a + d + (a + d)2 − 4
This is the normal form of the Möbius transformation. The constant K is called the
multiplier of the Möbius transformation. Observe that it depends solely on the sum
(a + d). The importance of this fact will become clear subsequently.
(ii) Finite, coincident fixed points: Recall that this case corresponds to c
= 0 and
(a + d) = ±2. The fixed point is given by ξ1 = ξ2 = ξ = (a − d)/(2c). The multiplier
reduces to K = 1. Starting with w = (az + b)/(cz + d), a little algebra leads to the
normal form in this case, which is
1 1
= ± c, depending on whether a + d = ±2.
w−ξ z−ξ
49
Again, this is the normal form in the present case. The multiplier K remains equal to
unity, of course.
4. Verify that the normal forms in the four cases above are as given above.
In its original form, the final expression for z (n) will obviously be quite complicated.
But if you use the normal form, you can write down the answer by inspection, in each
of the four cases listed above.
50
It is often useful to re-write the formula for z (n) in the general case (i) in the form
(ξ1 − K n ξ2 ) z + ξ1 ξ2 (K n − 1)
z (n) = .
(1 − K n ) z + (K n ξ1 − ξ2 )
This form shows explicitly that the result of an n-fold iteration of a Möbius trans-
formation is again such a transformation.11 The relation above is easily inverted to
express z ≡ z (0) in terms of z (n) . All you have to do is to replace K n by K −n . It is
easy to see that this merely entails an interchange of ξ1 and ξ2 . Thus,
(ξ2 − K n ξ1 ) z (n) + ξ1 ξ2 (K n − 1)
z (0) = .
(1 − K n ) z (n) + (K n ξ2 − ξ1 )
For any value of z, the set of points {z (n) |n ∈ Z} is an orbit under the map rep-
resented by a Möbius transformation. Such an orbit is analogous to the orbit of a
point in the phase space of a dynamical system in discrete time, n playing the role
of time. The collection of orbits corresponding to different values of z = z (0) (or the
initial conditions, in the case of a dynamical system) gives us a flow in the complex
plane Ĉ (analogous to a phase portrait in phase space). The nature of this flow is
essentially determined by the fixed points, which are the analogs of the critical points
or equilibrium points in phase space. This is the starting point of the classification of
Möbius transformations.
(i) If |K| > 1, the factor K n will grow in magnitude with increasing n. Therefore
z (n) → ξ2 as n → ∞, for all initial points z (0) (other than ξ1 , of course). In the spirit
of dynamical systems, we may regard ξ2 as an asymptotically stable fixed point, or
attractor; the flow is generally toward this point. On the other hand, ξ1 acts like an
11 th
Incidentally, this expression should
a b suggest to you the possibility of a simple ‘formula’ for the n
power of a general (2 × 2) matrix c d with unit determinant. Check this out.
51
unstable fixed point, or repellor; the flow is generally away from this point.
(ii) If |K| < 1, the factor K n tends to zero with increasing n. Hence z (n) → ξ1 as
n → ∞, for all initial points z (0) (other than ξ2 ). The attractor is now ξ1 , and the flow
is toward this point. The repellor is ξ2 , and the flow leads away from this point.
(iii) When |K| = 1, we have the analog of marginal fixed points, and the flow is neither
toward nor away from the fixed points.
(iv) The case when the two fixed points coincide at ξ = (a − d)/(2c) occurs when
(a + d)2 = 4, and corresponds to K = 1. This is a special case of |K| = 1.
The crucial point is that, in all cases, K depends only on the trace T ≡ (a + d) of the
matrix ( ac db ) made up of the coefficients of the transformation. The multiplier is given
in terms of T by √
T − T2 − 4
K= √ .
T + T2 − 4
52
This is the starting point for the classification of Möbius transformations into different
types. Four types of transformations are possible, depending solely on the value of
T . I merely list them here, without going into further detail. The first three types
correspond to real values of the trace T .
Type 1: Elliptic transformation, when T is real and −2 < T < 2. Then |K| = 1,
but K
= 1.
Type 2: Parabolic transformation, when T = ±2. The two fixed points now
coincide, and we have K = 1.
The isometric circle: As stated already, the local stretch or contraction factor
associated with the mapping of an infinitesimal neighborhood of any point z by the
Möbius transformation z → (az + b)/(cz + d) is |f (z)| = 1/|cz + d|2 . It follows that
there is no distortion at all in the map of the circle given by
Points inside the isometric circle satisfy |cz + d| < 1, and hence infinitesimal area
elements inside the circle are magnified by the mapping, by a factor |cz + d|−4 . Simi-
larly, points outside the isometric circle satisfy |cz + d| > 1. Hence infinitesimal area
elements in this region are shrunk by the transformation, by a factor |cz + d|−4.
53
(respectively, exterior) of the isometric circle in the z-plane is mapped to the
exterior (respectively, interior) of its image in the w-plane. (As you know, the
center −d/c of the isometric circle is mapped to ∞, while ∞ is mapped to the
center a/c of the image of the isometric circle under the mapping.)
(b) Consider the case of finite, distinct fixed points ξ1 and ξ2 , and (i) |K| > 1 and
(ii) |K| < 1, respectively. What is the isometric circle of the nth iterate of the
transformation? What happens as n → ∞?
Möbius transformations are easily seen to satisfy the axioms that define a group:
Thus, if
a1 z + b1 a2 z + b2
z → z = is followed by z → z = ,
c1 z + d 1 c2 z + d 2
then
(a2 a1 + b2 c1 )z + a2 b1 + b2 d1
z = .
(c2 a1 + d2 c1 )z + c2 b1 + d2 d1
But the set of all nonsingular (2 × 2) matrices with complex elements constitutes a
group. The group composition law is just matrix multiplication. The group is called
the general linear group over the complex numbers in 2 dimensions, and is denoted
by GL(2, C). As you know, however, the coefficients of a Möbius transformation can
12
That is, if w = f (z) = (az + b)/(cz + d), then z = f −1 (w) = (dw − b)/(−cw + a).
54
always be chosen such that ad − bc = 1. What is relevant here, therefore, is the set of
(2×2) matrices with complex elements and determinant equal to unity. These matrices
form a subgroup of GL(2, C): the unimodular or special linear group over the complex
numbers, denoted by SL(2, C). We might therefore expect Möb (2, C) to be essentially
the same as SL(2, C). But there is one more point to be taken into account.
Changing the sign of all the four coefficients in a Möbius transformation does not
alter it, because (az + b)/(cz + d) ≡ (−az − b)/(−cz − d). Hence, to each Möbius
transformation
az + b
z → w = with ad − bc = 1,
cz + d
there correspond the two matrices
a b −a −b
and ∈ SL(2, C).
c d −c −d
• The matrices I and −I form the kernel of the homomorphism. That is, {I, −I}
is the set of matrices in SL(2, C) whose image in Möb (2, C) is the identity trans-
formation.
But the two matrices I and −I themselves form a group, under matrix multiplication.
This group is just the cyclic group of order 2, denoted by Z2 . It is the same as the
group formed by the integers 0 and 1 under binary addition, i.e., addition modulo 2.
One now forms the quotient group SL(2, C)/Z2 . This is read as ‘SL(2, C) modulo Z2 ’,
and refers, in effect, to the group of unimodular 2 × 2 matrices up to the overall sign of
the matrices. In other words, it is a group of these matrices, such that any two matrices
differing only in sign are identified with each other and regarded as a single element of
the group. It is this group, also known as the projective special linear group P SL(2, C),
with which the group of Möbius transformations is in one-to-one correspondence.
Möb (2, C) ∼
= SL(2, C)/Z2 ≡ P SL(2, C).
• The group SL(2, C) is the so-called universal covering group of Möb (2, C).
The latter is a subgroup of SL(2, C).
Finally, I mention the following remarkable fact that connects Möbius transformations
to special relativity. The special linear group SL(2, C) is also the so-called universal
55
covering group of the group of homogeneous, proper, orthochronous Lorentz trans-
formations, SO(3, 1), in the usual spacetime comprising three spatial dimensions and
one time dimension. There is again a 2-to-1 homomorphism between SL(2, C) and
SO(3, 1), and we have
Möb (2, C) ∼
= SL(2, C)/Z2 ∼
= SO(3, 1).
That is, the group of Möbius transformations is isomorphic to the Lorentz group!
The Möbius group over the reals: We’ve seen that the Möbius group Möb (2, C)
is isomorphic to certain other important groups such as the projective linear group
P SL(2, C) and the homogeneous Lorentz group SO(3, 1). Further, it is a subgroup of
the special linear group SL(2, C), and hence of the general linear group GL(2, C). In
turn, the Möbius group itself has some important and interesting subgroups.
• a real Möbius transformation maps the upper (respectively, lower) half of the
complex plane to the upper (respectively, lower) half-plane.
The modular group: Among Möbius transformations with real parameters (a, b, c, d)
where ad − bc = 1, those with integer values of the parameters form a group by them-
selves! This is the group P SL(2, Z), called the modular group Similarly, the set of
Möbius transformations in which (a, b, c, d) are Gaussian integers, i.e., each parameter
is of the form m + ni where m and n are integers, also forms a group. The modular
group has numerous remarkable properties, and it plays a role in many areas of math-
ematics including hyperbolic geometry, number theory, elliptic functions, etc. It also
appears in certain topics in theoretical physics such as conformal field theory.
The invariance group of the unit circle: Here is another important example of a
continuous subgroup of the Möbius group. Consider all Möbius transformations that
56
leave the unit circle unchanged, i.e., those that map the circle |z| = 1 in the z-plane
to the circle |w| = 1 in the w-plane. These transformations are either of the form
az + b a b
z → w = ∗ , where det ∗ ∗ = |a|2 − |b|2 = 1,
b z + a∗ b a
or of the form
az + b a b
z→
w= , where det = −|a|2 + |b|2 = 1.
−b∗ z − a∗ −b −a∗
∗
Let’s call these type (i) and type (ii) transformations, for convenience. They form a
subgroup of the group Möb (2, C) of Möbius transformations. Transformations of types
(i) and (ii) are distinguished by the following property:
— A type (i) transformation maps the interior of the unit circle in the z-plane to
its interior in the w-plane; and the exterior of the unit circle in the z-plane to
the exterior of the unit circle in the w-plane.
— A type (ii) transformation maps the interior of the unit circle in the z-plane to
its exterior in the w-plane, and vice versa.
8. Show that
(a) the general form of a Möbius transformation z → w that maps the unit circle to
the unit circle is as given above;
(b) all such transformations form a group;
(c) |z| ≶ 1 ⇒ |w| ≶ 1 for type (i) transformations, and |z| ≶ 1 ⇒ |w| ≷ 1 for type
(ii) transformations.
Connection with the pseudo-unitary group SU(1, 1): The matrices correspond-
ing to Möbius transformations of types (i) and (ii) above are, respectively, of the general
form
a b
U+ = ∗ ∗ where det U+ = |a|2 − |b|2 = 1,
b a
and
a b
U− = where det U− = −|a|2 + |b|2 = 1.
−b −a∗
∗
But these are precisely the general forms of the matrices belonging to the indefinite
unitary group (or pseudo-unitary group) SU(1, 1). This is the group of (2 × 2)
matrices with complex entries satisfying the following conditions: if U ∈ SU(1, 1), then
† 1 0
det U = +1 and U g U = g, where g = .
0 −1
57
Now, SU(1, 1) turns out to be isomorphic to the special linear group SL(2, R) of
unimodular (2 × 2) matrices over the real numbers. Further, the group SL(2, R) is
also isomorphic to the symplectic group Sp(2, R), which is the group of canonical
transformations of a Hamiltonian system with one degree of freedom! We have
SU(1, 1) ∼
= SL(2, R) ∼
= Sp(2, R),
and the group of Möbius transformations that leaves the unit circle invariant is there-
fore isomorphic to SL(2, R)/Z2 , namely, Möb (2, R) once again.
The transformation e3 can also be regarded as successive reflections about the line
Re z = 12 and about the real axis, performed in either order. Let ei ej stand for the
transformation ej followed by the transformation ei . It is obvious that e2 e2 = e1 =
e3 e3 , so that e−1
2 = e2 and e−13 = e3 . Further, e4 ≡ e3 e2 and e5 ≡ e2 e3 . The
last element e6 = e2 e3 e2 (inversion-rotation-inversion) or, alternatively, e6 = e3 e2 e3
(rotation-inversion-rotation).
58
Note that this group is not an abelian or commutative group, since ei ej
= ej ei in
general. Several identities follow at once from the multiplication table above: for
instance, (e4 )3 , (e5 )3 , (e6 )2 , (e3 e4 )2 , (e4 e3 )2 , (e3 e5 )2 , (e5 e3 )2 , . . . are all equal to the
identity transformation.
59
6 Multivalued functions; integral representations
Branch points and branch cuts: An analytic function f (z) = w should be regarded
as a map f : z → w of some region R of the extended complex plane Ĉ to some region
R of Ĉ. If f (z) is single-valued, this map yields a unique value of w for every value
of z. All the functions we have considered so far satisfy this property. In general,
however, we have to deal with multivalued functions.
A simple example is provided by the function f (z) = z 1/2 . It is obvious that the
whole of the complex z-plane is mapped to just the upper half-plane in w, because
z = r eiθ maps to w = r 1/2 eiθ/2 . In order to cover the whole of the w-plane, the
z-plane has to be covered twice: the argument θ of z has to run from √ 0 to 4π, rather
than just 2π. As θ increases from 0 to 2π, we obtain the branch + z of the function √
f (z) = z 1/2 . As it increases further from 2π to 4π, we obtain the branch eiπ z 1/2 = − z
of the square root function.
In order to keep track of the correct branch, we therefore need two copies of the
z-plane. The two copies may be imagined to be two sheets, one lying below the other,
such that we descend to the second sheet by starting just above the positive real axis
on the first sheet and traversing a path that encircles the origin once in the positive
sense. On encircling 0 once again, this time on the second sheet, we ascend back to
the first sheet. The two sheets (labelled I and II, say) are called Riemann sheets,
and they are supposed to be connected to each other along a slit in each sheet running
from 0 to ∞ on the positive real axis. The top sheet, on which the phase of z runs from
0 to 2π, is called the principal sheet of the function concerned. The points where
the two branches of the two-valued function z 1/2 coincide in value, namely, z = 0 and
z = ∞ (recall that there is only one point at infinity in Ĉ), are called branch points.
They are connected by a branch cut. The two sheets pasted together as described
above form the Riemann surface of the function z 1/2 .
The branch cut joining z = 0 and z = ∞ may be taken to run along any curve
running between these two points. It is most convenient to take it to run along a
straight line. We have chosen the positive real axis in the foregoing. But this need not
always be so. All that is needed is a specification of the phase (or argument) of the
function concerned just above and just below the branch cut, so that we can calculate
the discontinuity (or the jump in the value) of the function across the branch cut.
Label the function on the sheets I and II as fI (z) and fII (z), respectively. By continuity,
the value of the function on sheet I, as you approach the positive real axis from below,
is the same as the value of the function on sheet II, as the real axis is approached from
above. That is,
lim fI (x − i) = lim fII (x + i), x > 0.
→0
→0
The discontinuity across the cut is then easily determined. In the present instance, it
60
is given by
disc f (z) = lim [fI (x + i) − fI (x − i)]
x>0
→0
Winding point: The function z α , where α is not a rational real number, also has
branch points at z = 0 and z = ∞. These are called winding points. The Riemann
surface of this function has an infinite number of sheets, because e2πniα is never equal
to unity for any integer value of n. These sheets are labelled by an integer n ∈ Z. On
the principal sheet, the phase of z runs from 0 to 2π, as usual, and n = 0.
Branch cuts run from one branch point to another. No function can have just one
branch point; the smallest number of branch points that a function can have is two.
When a branch point is encircled in the z-plane, the function does not return to its
original value. In order to obtain a closed contour that encircles a branch point, the
contour must encircle the branch point (or other branch points) as many times as is
necessary to return the function to its original value at the starting point.
Consider the functions (z − a)1/2 (z − b)1/2 and (z − a)1/2 /(z − b)1/2 , where a and b
are any two finite complex numbers. They have algebraic (square-root) branch points
at z = a and z = b, but their behavior at z = ∞ is regular. Therefore their branch
cuts can be chosen to run over the finite segments from a to b.
For any arbitrary non-integer value of α, including complex values, the cuts of the
function (z − a)α /(z − b)α can also be chosen to run over the finite segment from z = a
to z = b alone. But when α is not a half-odd-integer, this is not possible for the product
function (z − a)α (z − b)α . This function has branch points at z = a, z = b, as well as
z = ∞, for all α that is not an integer or half-odd-integer. The branch cut structure
61
of this function must necessarily run up to ∞.
The singularity structure of multivalued functions can be quite intricate and inter-
esting. Here’s a simple example. The function
1
f (z) = ln (1 − z)
z
has logarithmic branch points at z = 1 and z = ∞. On the principal sheet of the
logarithm, we have ln 1 = 0; hence ln (1 − z) −z in the infinitesimal neighborhood
of the origin on this sheet. The simple pole at z = 0 of the factor z −1 is therefore ‘can-
celled’ by the simple zero of the logarithm, and f (z) only has a removable singularity
at z = 0. On every other sheet of the logarithm, however, f (z) does have a simple pole
at z = 0, because ln 1 = 2πni
= 0 on all these sheets.
Contour integrals in the presence of branch points: As you have seen, the eval-
uation of integrals via contour integration relies, ultimately, on integrating analytic
functions over closed contours. In the case of multivalued functions, however, if the
contour starts at some point z and encircles a branch point before returning to the
starting point, the function does not return to its original value. Hence the contour is
not really a closed one—the final point is actually on some other Riemann sheet of the
function!
In order to apply the theorems pertaining to integrals over closed contours, you
must ensure that the function has returned to its starting value (or, equivalently, that
z has returned to the starting point on the Riemann surface of the function). In gen-
eral, this involves encircling more than one branch point, or the same branch point
more than once, and so on, as will become clear from the examples that follow.
1. Let a and b be arbitrary real numbers, where a < b. Use contour integration to show
that the integral b
dx
I= = 2π,
a (b − x)(x − a)
independent of a and b.
Hint: Consider the function f (z) = (z − a)−1/2 (z − b)−1/2 . Note that the behavior
of the function at ∞ is regular. Choose the branch cut to run from a to b, and write
down the phases of the function above and below the cut. Relate the line integral I to
a contour integral surrounding the branch cut. Now open out the contour to make it a
large circle whose radius tends to infinity, and pick up the contribution from the circle
to arrive at the result. This method makes it evident why the value of the integral is
independent of the actual values of a and b.
62
2. Use a similar method to show that
dz
√ = 2πi,
C 1 + z + z2
where C is the circle |z| = 2 traversed once in the positive sense.
Hint: The branch points of the integrand are at two of the cube roots of unity, given
by z = ω = e2πi/3 and z = ω 2 = e4πi/3 . The branch cut of the integrand may be taken
to run between these points. As |z| → ∞ along any direction, the integrand tends to
1/z.
(i) The degree of q(x) exceeds that of p(x) by at least 2. Hence the integrand decays
at least as rapidly as 1/x2 as x → ∞, and the convergence of the integral is
guaranteed.
(ii) q(x) does not have any zeroes for x ≥ 0. Hence there is no non-integrable
singularity on the path of integration, and the integral exists.
where C is the familiar hairpin contour that comes in from infinity just below the
positive real axis, encircles zero from the left, and runs just above the positive real
axis. The branch cut of ln z is taken to run from 0 to ∞ along the positive real axis
in the z-plane. Since ln z = ln x just above the cut, and ln z = ln x + 2πi just below
the cut, we have
0 ∞
1 p(x) (ln x + 2πi) 1 p(x) ln x
I =− dx − dx = I.
2πi ∞ q(x) 2πi 0 q(x)
13
Obviously, you must then take care to avoid the appearance of any spurious logarithmic singularity
owing to the upper limit of integration. Evaluate the integral up to some upper limit L, and take the
limit L → ∞ after the different terms are recombined properly.
63
On the other hand, we can evaluate I by completing the contour C with the addition
of a large circle (that does not cross the cut). The contribution of the large circle
vanishes as its radius R of the circle tends to infinity, because the integrand vanishes
at least as fast as (ln R)/R on this circle. But the closed contour can now be shrunk
to pick up the residues of all the poles of the integrand (located at the zeroes of q(z)).
The integral I is thus evaluated quite easily.
3. Use the method just described to evaluate the following standard integrals:
∞
dx π π
(a) n
= cosec (n = 2, 3, . . . ).
0 x +1 n n
∞
dx π 2π
(b) = cosec (n = 3, 4, . . . ).
0 xn−1 +xn−2 +···+ x+1 n n
Hint: (a) The poles of the integrand that are enclosed by C are at the roots of
z n + 1 = 0, namely, at z = eiπ/n ω r , where ω = e2πi/n and r = 0, 1, . . . , n − 1.
(b) The poles of the integrand that are enclosed by C are at the roots of unity other
than 1 itself, i.e., at z = ω r , where r = 1, . . . , n − 1.
Contour integral representations for various special functions are easily constructed
using the properties of multivalued functions. Such representations often provide ‘mas-
ter representations’ or analytic continuations of the functions concerned that are valid
for all complex values of the arguments of these functions. Let’s consider some exam-
ples.
The gamma function: Recall the defining representation of this function, namely,
∞
Γ(z) = dt tz−1 e−t , Re z > 0.
0
We’ve seen that the functional equation z Γ(z) = Γ(z + 1) helps us extend the region
of analyticity to the left of the region Re z > 0, stripwise. But we are now in a position
to write down a single representation for the gamma function that is valid throughout
the complex plane.
Consider the integrand tz−1 e−t in the defining representation as an analytic function
of t, for any arbitrary complex value of z. The factor tz−1 has a branch points at t = 0
and t = ∞, with a cut running between them. Take the branch cut to run along the
positive t-axis. The phase of tz−1 just above the cut is 0, while
it is 2πiz just below
−2πi z−1 −t
the cut (the factor e is just unity). Consider the integral C dt t e , where the
contour C comes in from ∞ to just below the cut, encircles the origin in the negative
sense in an arc of a circle of radius , and runs just above the cut from to ∞. As
long as Re z > 0, the contribution from the arc of the small circle vanishes as → 0.
64
Moreover, the contributions from the two line segments equal −e2πiz Γ(z) and Γ(z),
respectively. Hence, in the region Re z > 0, we have
1
Γ(z) = dt tz−1 e−t .
(1 − e2πiz ) C
But C does not pass through the point t = 0, and may be deformed to stay clear of
the origin, in the form of a hairpin-shaped contour straddling the branch cut of tz−1 .
The contour integral is thus defined for all finite values of z. On the other hand, the
factor (1 − e2πiz )−1 has simple poles at all integer values of z. The product of these
two factors is a meromorphic function of z. It represents Γ(z) for all z. Note that the
hairpin contour C can be partially ‘straightened out’, but we must always ensure that
Re t → +∞ asymptotically at both ends of the open contour, so that the damping
factor e−t ensures the convergence of the integral.
4. It can be checked that the known analytic properties of Γ(z) follow from the integral
representation derived above.
(a) Owing to the factor 1/(1 − e2πiz ), it might appear that Γ(z) has a simple pole
at every integer value of z. But we know that Γ(n) is finite (= (n − 1)!) when
n is a positive integer. What happens is that the contour integral C dt · · · also
vanishes when z = 1, 2, . . . : the integrand becomes single-valued because tn−1
does not have any branch points, and the two straight segments of the contour
cancel each other out. Verify that, as z tends to any positive integer n, the
limiting value of the right-hand side is precisely (n − 1)!.
(b) On the other hand, when z = −n where n = 0, −1, −2, . . . , the branch cut
disappears, but a pole of order (n + 1) is left at the origin in the t-plane. The
contour integral is then evaluated easily. Verify that Γ(z) has a simple pole at
z = −n with residue equal to (−1)n /n!.
The beta function: Recall the original definition of the beta function,
1
B(z, w) = dt tz−1 (1 − t)w−1 , where Re z > 0, Re w > 0.
0
You have also seen that, in this case, it is not possible to extend the region of analyt-
icity in z and w simultaneously by using integration by parts.
Once again, we can continue B(z, w) analytically to all values of z and w by suit-
ably exploiting the discontinuity of the integrand tz−1 (1 − t)w−1 across the branch cut
running between the branch points at t = 0 and t = 1. For general values of z and w,
there is a branch point at t = ∞ as well. The branch cut structure of the integrand
is therefore more complicated in this case, because the cut actually runs all the way
to infinity. But all that we need to keep track of is the following fact: starting on the
65
principal sheet, the integrand acquires a factor e2πiz (respectively, e−2πiz ) whenever
the contour encircles the branch point at t = 0 in the positive (respectively, negative)
sense. Similarly, a factor e2πiw or e−2πiw results when the branch point at t = 1 is
encircled in the positive or negative sense.
What is required in order to return to the original value of the function, and hence
to close the contour, is a double encircling of each of the branch points t = 0 and
t = 1, once in the positive sense and once in the negative sense. The outcome is a
Pochhammer contour that is written symbolically as C : (1−, 0−, 1+, 0+).14 The
contributions from the infinitesimal circles around 0 and 1 vanish as long as Re z > 0
and Re w > 0. Therefore
1
1
B(z, w) = dt t (1 − t)
z−1 w−1
= dt tz−1 (1 − t)w−1 .
0 (1 − e−2πiz )(1 − e−2πiw ) C
But the contour of integration C does not pass through the singularities of the integrand
at t = 0 and t = 1. The contour can be distorted away from these points without
changing the value of the integral. The contour integral thus provides an analytic
continuation of the beta function for all finite values of z and w:
1
B(z, w) = dt tz−1 (1 − t)w−1 .
(1 − e−2πiz )(1 − e−2πiw ) C
The Riemann zeta function: Recall that the Riemann zeta function is defined as
∞
1
ζ(z) = (Re z > 1).
n=1
nz
Once again, an analytic continuation of ζ(z) to the whole of the complex z-plane is
easily derived, in terms of a contour integral. Consider the product
∞ ∞
1
ζ(z) Γ(z) = z
du uz−1 e−u , where Re z > 1.
n=1
n 0
66
The divergence of the sum ∞ n=1 n
−z
that occurs when Re z = 1 now becomes a diver-
gence of the integral: near t = 0, the integrand behaves like tz−2 , an extra factor of t−1
coming from the denominator (et − 1)−1 . Hence the integral diverges unless Re z > 1.
5. The contour integral in the formula above is an entire function of z. It follows that
the only possible singularities of ζ(z) in the finite part of the complex plane must come
from the poles of the factor Γ(1 − z) at z = 1, 2, . . . .
(a) Show that ζ(z) has a simple pole at z = 1, with residue equal to 1.
(b) Show that ζ(z) has no singularities at z = n where n = 2, 3, . . . .
Hint: (a) At z = 1, the contour integral reduces to dt/t = −2πi.
67
• ζ(z) is a meromorphic function of z, with a simple pole at z = 1. The residue at
this pole is equal to unity.
Connection with Bernoulli numbers: We have seen earlier that ζ(2n), where n is
a positive integer, is π 2n multiplied by a rational number less than unity. I have also
mentioned there that no such simple closed-form expression exists for ζ(2n + 1). In
contrast, the value of ζ(z) at 0 and at the negative integers can be determined quite
easily, as follows.
The function 1/(et − 1) has a simple pole at t = 0, with residue equal to unity. The
function t/(et − 1) is analytic in the neighborhood of the origin.15 Its Taylor expansion
about t = 0 is
t ∞
tn
= Bn ,
(et − 1) n=0 n!
where the constants Bn are certain rational numbers, called the Bernoulli numbers.
These are defined by the expansion above. Thus t/(et − 1) is the generating function
for the Bernoulli numbers. The first few numbers are found to be
B0 = 1, B1 = − 12 , B2 = 16 , B4 = − 30
1
, B6 = 1
42
, ... .
Interestingly, all the odd Bernoulli numbers B2n+1 are equal to zero, except for B1 .
The sequence |B2 |, |B4 |, |B6 |, . . . appears to be a decreasing sequence. You might
therefore think that the series ∞ n
n=0 Bn t /n! ought to converge at least as well as the
series for the exponential et . The radius of convergence would then be infinite, and
the series would represent an entire function. But t/(et − 1), the function represented
by the series, has poles at t = 2πni, where n = ±1, ±2, . . . . Hence the radius of
convergence of the series must be the distance from the origin to the nearest of these
singularities, i.e., 2π. After the first few numbers, the Bernoulli numbers B2n actually
start increasing in magnitude quite rapidly with increasing n.
6. Use the expansion above in the representation of the zeta function to establish the
following results:
(a) ζ(0) = − 12 .
68
Hint: When z is zero or a negative integer, the factor tz−1 in the integrand does not
have any branch points. Instead, there is a pole at t = 0. The contour C then collapses
to a small circle encircling the pole once in the clockwise sense. The integral is easily
evaluated by the residue theorem.
The Riemann Hypothesis: The zeroes of ζ(z) at even negative integer values of z
are called the trivial zeroes of the zeta function. ζ(z) also has an infinite number of
other zeroes in the strip 0 < Re z < 1.
• The famous Riemann hypothesis asserts that all of these nontrivial zeroes lie
on the so-called critical line Re z = 12 .
It is known that an infinite number of zeroes do lie on the critical line. It is also known
that all nontrivial zeroes lie on that line ‘almost surely’ (in the sense of probability
theory, with probability equal to 1). The first 13 billion or so zeroes have indeed been
verified explicitly to lie on the critical line.16 The Riemann conjecture is perhaps the
most important unsolved problem in mathematics. It has resisted a rigorous and com-
plete proof for over a century and a half.17 A very large number of other results in
mathematics rest on the validity of the hypothesis. There are also several intriguing
connections between the distribution of the zeros of ζ(z) on the critical line, on the
one hand, and physical problems, on the other—for instance, the level-spacing of the
energy eigenvalues of quantum mechanical systems whose classical counterparts are
chaotic, of the eigenvalues of certain classes of random matrices, and of the energy
eigenvalues of complex nuclei. It is clear that the zeta function and its counterparts
and analogs in classical and quantum dynamical systems hide many secrets yet to be
discovered.
The Legendre functions Pν (z) and Qν (z): As another example of integral repre-
sentations of special functions, let’s consider the Legendre functions of the first
and second kinds, Pν (z) and Qν (z), for complex z and ν. These functions appear
in the solution of a very large number of physical problems. They are the linearly
independent solutions of Legendre’s differential equation,
2
2 d d
(1 − z ) 2 − 2z + ν(ν + 1) φ(z) = 0.
dz dz
They turn out to have the contour integral representations given below, for complex
values of both the argument z and the order ν.
Consider the function f (t ; z, ν) = [(t2 − 1)/2(t − z)]ν of the complex variable t, for
general complex values of z and ν. There are branch points (winding points, in general)
16
ζ(z) can be shown to be a real analytic function. Therefore if ζ( 12 + iy) = 0 for any real value of
y, it follows that ζ( 12 − iy) = 0.
17
You might want to give it a try!
69
at t = 1, t = −1 and t = z. Suppose we start at some point on the principal sheet of
f (t; z, ν), and move in a path that encircles one or more of the branch points. In how
many different ways can this be done, such that the function returns to its original
value when we return to the starting point? In each case, the path then becomes a
closed contour over which f (t; z, ν) can be integrated.
The Legendre functions of the first and second kind are then given by the formulas
2 ν
1 dt (t − 1)
Pν (z) =
2πi C1 (t − z) 2(t − z)
and
2 ν
1 dt (t − 1)
Qν (z) = .
π(e2πνi − 1) C2 (t − z) 2(t − z)
The analytic properties of Pν (z) and Qν (z) can be deduced from these representations.
Of the extensive set of these properties, I mention the following:
Pν (z) for non-integral order ν: For general non-integer values of the order ν, the
function Pν (z) is no longer a polynomial in z. It has branch points at z = −1 and
z = ∞. It is conventional to choose the branch cut to run from −1 to −∞ on the real
axis (the x-axis) in the z-plane. The discontinuity across the cut at any any point x is
again proportional to Pν (|x|) itself.
Qν (z) for non-integral order ν: Similarly, for general non-integer values of the
order ν, the function Qν (z) has branch points at z = 1, z = −1 and z = ∞. It is
conventional to choose the branch cut to run from 1 to −∞ on the x-axis. The dis-
continuity of Qν (x) for x ∈ (−1, 1) is proportional to Pν (x), while the discontinuity at
x < −1 is proportional to Qν (|x|). Based on these properties, it is possible to write
down dispersion relations for the functions Pν (x) and Qν (x).
70
Pl (z) for integer values of the order l: Consider Pν (z) when ν = l, where
l = 0, 1, . . . . The contour C1 now encloses no branch points, but only a pole of
order (l + 1) at t = z. The integral is then evaluated easily, and the Rodrigues
formula for Pl (z) is recovered:
1 dl 2
l 1 dl 2
Pl (z) = l (t − 1) ≡ (z − 1)l .
2 l! dtl t=z 2 l l! dz l
You would have encountered this formula in the case when z is a real variable lying
in the range [−1, 1]. We see now that this formula is also valid for complex values of
z. Note that Pl (z) continues to remain a polynomial of order l in the complex variable z.
When ν is a negative integer −(l + 1), the contour C1 encloses a pole of order (l + 1)
at t = 1. Once again, the integral can be evaluated. The symmetry property
can then be deduced. This is just a special case of the more general reflection symmetry
where Rl−1 is a polynomial of order l−1. There is a simple and very useful formula that
connects Ql (z), for any arbitrary complex value of the argument z, to the Legendre
polynomial of order l. It is
1 1 Pl (t)
Ql (z) = dt , l = 0, 1, . . . .
2 −1 (z − t)
71
When ν is a negative integer −(l + 1), the contour C2 encloses poles of order (l + 1)
at z = −1 and z = 1. The contour integral makes a finite, nonzero contribution. The
factor (e2πνi −1) in the denominator then leads to a simple pole of Qν (z) at ν = −(l+1).
The residue at the pole turns out to be Pl (z) itself.
Finally, I mention that such contour integral representations exist for all other
special functions as well. In most cases, they serve as analytic continuations of the
corresponding functions to the largest possible domain of their arguments, including
parameters such as the order, degree, and so on.
• How does a function defined by an integral become singular, and what are the
possible singularities?
As you might expect, the answer to such a general question involves whole branches of
mathematics (such as homology theory). Let’s narrow the question down very con-
siderably, and examine the simplest possibilities. The treatment below is elementary,
heuristic, and essentially based on a few simple examples.
where the path of integration is an open contour running from some point a to some
other point b in the complex t-plane. The integrand φ(t, z) is assumed to be analytic
in t and z, with some singularities. In general, when the integral exists, it defines an
analytic function of z. Let’s start with a value of z in the region in which f (z) is
holomorphic, and examine how the function could develop a singularity as z moves out
of this region. This is determined by the behavior of the singularities in the t-plane of
the integrand φ(t, z).
These singularities are of two kinds: they could be fixed or z-independent, such as
a pole at some point t = c (not lying on the original path of integration); or they could
be movable, or z-dependent, such as a pole at t = z. As z changes, one or more of
the latter could approach the path of integration in the t-plane. If the contour can be
distorted away so as to avoid the singularity (keeping it pinned down at the end-points
a and b, of course), we have an analytic continuation of f (z), defined by the integral
over the distorted contour. There are two cases in which this simple device will not
work. Each of them leads to a singularity of f (z).
72
(i) End-point singularity: If a moving singularity approaches one of the end-points
of integration, a or b, the contour cannot be moved away and an end-point singu-
larity of f (z) ensues. Consider the (extremely!) elementary example
1
dt
f (z) = .
0 t−z
The integrand has a moving pole at t = z. As it stands, the integral exists for all
z
∈ [−1, 1]. If z approaches the real axis in the t-plane either from above or from below,
the contour of integration can be moved away ahead of the pole, and the integral will
continue to exist in each case.18 As z → 0 or 1 (in the z-plane), the pole in the t-
plane approaches one of the end-points of the contour, and a singularity of f (z) occurs.
This is corroborated by the explicit form for f (z) obtained by carrying out the trivial
integral above, to get z − 1
f (z) = ln .
z
Clearly, f (z) has logarithmic branch points at z = 0 and z = 1, confirming our expec-
tation.
1
7. Apply this result to the formula Ql (z) = 1
2 −1
dt Pl (t)/(z − t), to show that
18
As you might guess, however, the analytic continuations of f (z) will differ in the two cases,
suggesting already that we’re dealing with different branches of f (z).
73
(a) the Legendre function Ql (z) (where l = 0, 1, . . . ) has logarithmic branch points
at z = ±1 (as stated earlier);
(b) the discontinuity across the branch cut running from −1 to 1 on the real axis is
given by
% &
disc Ql (x) = lim Ql (x + i) − Ql (x − i) = −iπ Pl (x).
−1<x<1
↓0
(ii) Pinch singularity: If two moving singularities approach the same point on
the path of integration, but from opposite sides of the path, the contour gets trapped
between them, and cannot ‘escape’. A pinch singularity of f (z) may then occur.
The same thing happens if a moving singularity traps the contour between itself and a
fixed singularity of the integrand lying on the other side of the contour. Consider the
function 1
dt
f (z) = .
−1 (z − t )
2 2
74
t = −z. End-point singularities may be expected at z = ±1. We may expect these to
be ‘mild’ singularities, since (t ± z)−1/2 is an integrable singularity. (It is trivially seen
that f (z) does not diverge at z = ±1, but rather has the finite value π.) Further, as
z → 0, the contour of integration is pinched between these branch points. Hence z = 0
must also be a singularity of f (z). Note that if we simply set z = 0 in the expression
for f (z), the integral diverges.
In order to find f (z) explicitly, let’s start with z on the positive real axis, z = x > 1.
Then a simple change of the variable of integration yields
This is a multivalued function, and you must be careful about its branch structure
when writing down its analytic continuation to the rest of the z-plane. It is convenient
to re-write the arcsine function as a logarithmic function, using the identity
% &
sin−1 u = −i ln iu ± (1 − u2 )1/2 .
In order to choose the right sign before the radical, note the following: When 1 ≤
x < +∞, the integral representing f (x) is real positive; and it decreases monotonically
from π to 0 as x → ∞. It follows that
x
f (x) = 2i ln √ , 1 ≤ x < ∞.
i + x2 − 1
In this form, the function is (trivially) analytically continued to
z
f (z) = 2i ln √ .
i + z2 − 1
It is now easy to see that f (z) has
(i) a logarithmic branch point at 0, as well as
75
• As z → ±1, the contour C2 gets pinched between the singularities of the inte-
grand at t = z and at t = ±1, respectively. Hence Qν (z) has a singularities at
z = 1 and z = −1. The branch cut of Qν (z) is customarily taken to run from 1
through −1 to −∞ along the real axis.
8. With a little effort, you can show that the discontinuities of Pν (z) and Qν (z) across
their branch cuts on the real axis are as follows:
disc Pν (x) = 2i sin (πν) Pν (−x) for −∞ < x < −1
disc Qν (x) = 2i sin (πν) Qν (−x) for −∞ < x < −1
disc Qν (x) = −iπ Pν (x) for −1 < x < 1.
Hint: In order to find the discontinuity of Pν (z) across the cut from −1 to −∞, con-
sider its integral representation in the respective cases when z = x + i and z = x − i,
where x < −1. Note the configuration of the contour C1 in the two cases, along
with the branch cuts of the integrand. Write down the contour integrals segment by
segment, keeping careful track of the phases of the various factors in the integrand.
Compare the result with the contour integral for Pν (−x) (where −x > 1), to arrive
at the result quoted. A similar procedure will yield the corresponding results for Qν (x).
Dispersion relations for the Legendre functions: Finally, based on the disconti-
nuities found above, we can derive dispersions relations for the Legendre functions of
both kinds. As you might expect, the asymptotic (|z| → ∞) behaviors of Pν (z) and
Qν (z) are also required for this purpose. I do not go into this here, but merely quote
the relations, for completeness:
sin πν ∞ Pν (t)
Pν (z) = dt (ν
= integer),
π 1 (z + t)
and
1 1 Pν (t) sin πν ∞ Qν (t)
Qν (z) = dt + dt .
2 −1 (z − t) π 1 (z + t)
1
The relation Ql (z) = 12 −1 dt Pl (t)/(z − t) when ν = l (= 0, 1, . . . ) is a special case of
the last formula above.
76
7 Laplace transforms
In physical applications, we are often concerned with functions that are only defined
on a half-line, say [0, ∞). An example is a causal response function φ(t), where t
denotes the time variable. More generally, the whole class of initial value problems
involves functions of this kind. Such functions may vanish, or tend to a constant, or
even diverge as t → ∞. It is very helpful to define an integral transform of the
functions concerned, that turns differentiation with respect to t into multiplication by
the variable ‘conjugate’ to t.
The Laplace transforms of simple functions are easily written down. Consider, for
instance, the function
f (t) = tα e−at .
It is clear that its Laplace transform
∞
f(s) = dt tα e−(s+a)t
0
exists as long as Re α > −1 (so that the lower limit of integration, t = 0, does not
pose any problem), and also Re s > −Re a (so that there is no divergence owing to the
upper limit of integration, t = ∞). The integral is easily seen to be a gamma function.
We have
Γ(α + 1)
f (t) = tα e−at (Re α > −1) =⇒ f(s) = .
(s + a)α+1
20
Note that there do exist functions which do not possess a Laplace transform—e.g., if f (t) ∼
exp (t1+α ) (where α > 0) as t → ∞, then f(s) does not exist for any value of s.
77
f(s) may now be continued analytically to all values of s, using the explicit representa-
tion above. Observe that f(s) has a singularity at s = −a, in accord with the general
statement made earlier.
A number of simpler cases may be read off from this result. For instance,
1 n! s a
L[e−at ] = , L[tn ] = n+1 , L[cos (at)] = 2 2
, L[sin (at)] = 2 .
(s + a) s s +a s + a2
In general, if L[f (t)] = f(s), then L[f (t) e−at ] = f(s + a).
78
Hence show that
L[f (n) (t)] = sn f(s) − sn−1 f (0) − sn−2 f (1) (0) − . . . − s f (n−2) (0) − f (n−1) (0)
n
n
= s f (s) − sn−j f (j−1) (0).
j=1
The inverse transform: The inverse Laplace transform that expresses f (t) in terms
of f(s) is given by the Mellin formula
c+i∞
1
f (t) = ds est f(s),
2πi c−i∞
where the contour runs parallel to the imaginary axis in the s-plane, cutting the real
axis at a point c such that the contour stays to the right of all the singularities of f(s).
Hence the contour lies entirely in a region in which f(s) is analytic, as required to make
the integral well-defined. (Recall that f(s) generally has singularities in a certain left
half-plane.) Using the Mellin formula, it is easy to invert a Laplace transform whenever
f(s) is a rational function of s. The contour over s can then be closed by adding a
large semi-circle in the left half-plane, because the contribution of the semi-circle to
the integral vanishes as its radius tends to infinity: the factor est in the Mellin formula
ensures this. (Remember that t ≥ 0.) The closed contour now encircles all the poles
of f(s). The residue theorem then gives the value of the contour integral, and hence
f (t) is determined.
5. The Bessel function of the first kind and of order ν may be defined by the power
series
∞
(−1)n 1 2n+ν
Jν (z) = 2
z .
n=0
Γ(ν + n + 1) n!
The order ν does not necessarily have to be an integer, in this definition. Jν (z) is an
entire function of z. When ν = 0, we have
∞
(−1)n 1 2n
J0 (z) = 2 2
z .
n=0
(n!)
79
Now consider the function f(s) = (s2 +a2 )−1/2 , where a is a real constant (say). Expand
it in a binomial series in inverse powers of s, and invert the transform term by term
using the fact that the inverse transform of 1/s2n+1 is t2n /(2n)! Hence show that
−1 1
L √ = J0 (at).
s2 + a2
Since ∞
1
dt e−st J0 (at) = √ ,
0 s2 + a2
∞
it follows that 0
dt J0 (at) = 1/a for a > 0.
6. The iterate of the Laplace transform: Show that, provided the integrals
concerned exist,
∞ ∞ ∞
−us −st f (t)
L [f (t)] =
2
ds e dt e f (t) = dt .
0 0 0 (t + u)
LCR series circuit: As another standard example from elementary physics, consider
an LCR series circuit under a sinusoidal applied voltage of amplitude V0 and (angular)
frequency ω. Recall that we have already determined the complex admittance of the
system, when discussing linear response and dispersions relations.
As you know, this equation is precisely the equation of motion of a sinusoidally forced
damped simple harmonic oscillator. The damping constant is R/L = γ, which is just
the reciprocal of the time constant
√ of an LR circuit. The natural frequency of the circuit
in the absence of the resistor 1/ LC = ω0 . The condition ω0 > 12 γ corresponds to the
underdamped case. Let’s consider this case first , for definiteness.21 It is convenient,
in this case, to work in terms of the shifted frequency
1/2
ωu = ω02 − 14 γ 2 .
7. Suppose the initial conditions are such that both the initial charge and the initial
current are equal to zero, i.e.,
80
The solution for q(t) can be written as the sum of a transient part and a steady state
part,
q(t) = q tr (t) + q st (t).
Show that these are given, respectively, by
2
tr V0 ω (ω − ω02 + 12 γ 2 ) sin (ωu t) + ωu γ cos (ωu t) −γt/2
q (t) = e
L ωu (ω 2 − ω02 )2 + ω 2γ 2
and
V0 (ω 2 − ω02 ) sin (ωt) + ω γ cos (ωt)
q (t) = −
st
.
L (ω 2 − ω02 )2 + ω 2 γ 2
Hint: Take the Laplace transform of both sides of the differential equation. With the
initial conditions q(0) = 0 and q̇(0) = 0, we get
(V0 ω/L)
q(s) = .
(s2 + ω )(s2 + γs
2 + ω02)
Resolve the right-hand side into partial fractions, and invert the Laplace transform.
All you need to use is the fact that the inverse Laplace transform of (s + a)−1 is e−at .
Observe that the transient component of the solution, q tr (t), is characterized by the
frequency ωu that depends on the circuit parameters L, C and R. This part decays to
zero exponentially in time, owing to the dissipation in the system. The steady state
component q st (t), on the other hand, oscillates with the same frequency ω as the ap-
plied voltage. These statements apply equally to the current in the circuit, given by
I(t) = q̇(t).
8. The remarks just made are, in fact, applicable to all initial conditions. In order to
check this out, consider general values q(0) and q̇(0) = I(0), respectively, of the initial
charge on the capacitor and the initial current in the circuit. Show that the complete
solution for q(t) is now given by the one already found above, plus an extra term added
to q tr (t), namely,
I(0) + 12 γ q(0)
sin (ωu t) + q(0) cos (ωu t) e−γt/2 .
ωu
81
• In the solution of an inhomogeneous differential equation, the purpose of adding
‘the right amount’ of the complementary function to the particular integral is
to ensure that the boundary conditions (in this case, the initial conditions) are
satisfied by the solution.
The example just considered ought to help you see this quite clearly.
9. By now, it should be obvious to you that the solutions in the critically damped
(ω0 = 12 γ) and overdamped (ω0 < 12 γ) cases may be written down by simple analytic
continuation of the solution in the underdamped case.22 Do so.
Laplace transforms and random processes : The master equations for certain
Markov processes lead to simple differential equations for the generating functions of
these processes. Such equations are solved very easily using Laplace transforms. Here
are a few examples.
10. The Poisson process: Radioactive decay of an unstable isotope provides a phys-
ical example of a Poisson process. If Pn (t) is the probability that exactly n events of
the process take place in a time interval t, and λ > 0 is the mean rate at which events
take place, then
∂f
+ λ(1 − z) f = λz P0 = λz e−λt ,
∂t
with the initial condition f (z, 0) = 0. Use Laplace transforms to obtain the solution
f (z, t) = e−λt eλzt − 1 .
Picking out the coefficient of z n in the power series expansion of the exponential we
get the Poisson distribution
e−λt (λt)n
Pn (t) = .
n!
The mean number of events in a time interval t is therefore λt. Every cumulant of the
distribution is also equal to λt.
22
Typically, trigonometric functions will become hyperbolic functions.
23
Note that the generating function has been defined as a sum from n = 1 rather than n = 0.
82
The Poisson process is an example of what is known as a birth process, because
the random variable n(t) never decreases as t increases. The next example shows what
happens when the value of the random variable can both increase as well as decrease
as time elapses (a birth-and-death process).
11. A biased random walk on a linear lattice : Consider a linear lattice, with its
sites labelled by an integer j ∈ Z. A walker on this lattice jumps from any site to one
of the two neighboring sites with a mean jump rate λ. The probability of a jump from
the site j to the site (j + 1) is p, and that of a jump to the site (j − 1) is q = (1 − p),
where 0 < p < 1. Successive jumps are statistically independent of each other. Let
Pj (t) denote the probability that the walker is at site j at time t. (The random variable
in this problem is j). The master equation satisfied by the set of probabilities {Pj (t)}
is
dPj (t) % &
= λ p Pj−1(t) + q Pj+1(t) − Pj (t) , j ∈ Z.
dt
‘gain’ terms ‘loss’ term
Without loss of generality,24 we may take the site 0 to be the starting point of the
random walk (RW for short). The initial condition is then
Pj (0) = δj,0 .
Note that the summation is over all integers j. Hence the equation above is to be
regarded as a Laurent series expansion of f (z, t).
(a) From the differential equation satisfied by Pj (t) and its initial condition, show
that f (z, t) satisfies the equation
∂f
= λ (pz − 1 + qz −1 ) f ,
∂t
with the initial condition f (z, 0) = 1 .
83
(c) The probability we seek, Pj (t), is the coefficient of z j in the Laurent expansion
of f (z, t) in powers of z. It is helpful to write
−1 √ −1
pz + qz = pq z p/q + z q/p .
Now expand the second exponential factor in the solution for f (z, t), and pick
out the coefficient of z j . Consider j ≥ 0 first. Show that
∞
1 √
−λt
Pj (t) = e (p/q) j/2
(λt pq)2n+j , j ≥ 0.
n=0
n! (n + j)!
The modified Bessel function of the first kind and of order ν may be defined by
means of its power series, namely,
∞
1 1 2n+ν
Iν (z) = z .
n=0
Γ(ν + n + 1) n! 2
Compare this series with that for the ordinary Bessel function Jν (z) given earlier.
Apart from the factor (−1)n in the summand in the case of Jν (z), the series are the
same.25 Like the Bessel function Jν (z), the modified Bessel function Iν (z) is also an
entire function of z. The generating function for the modified Bessel function of integer
order is26
∞
t (z+z −1 )/2
e = Ij (t) z j .
j=−∞
(d) Compare the series expansion obtained above for Pj (t) with the power series for
the modified Bessel function, to obtain
√
Pj (t) = e−λt (p/q)j/2 Ij (2λt pq) .
Verify that exactly the same expression is valid for negative integer values of j
as well, using the fact that I−j = Ij for any integer j.
25
The two functions are related to each other according to Jν (z) = eiπν/2 Iν (−iz), or Iν (z) =
−iπν/2
e Jν (iz).
26
For completeness, I mention that the generating function of the Bessel function Jk is given by
∞
−1
et (z−z )/2
= Jj (t) z j .
j=−∞
84
An RW in which p
= q is called a biased random walk. When p > q [respectively,
q > p], there is a bias to the right [left], and the factor (p/q)j/2 in the expression for
Pj (t) shows that the probability of positive [negative] values of j is enhanced, at any
t > 0.
Since Ij = I−j , we have in this case Pj (t) = P−j (t) at any time t, as expected on
physical grounds.
eλt
Ij (λt) ∼ √ ,
2πλt
independent of j. Therefore, in an unbiased RW, Pj (t) has a leading asymptotic time-
dependence ∼ t−1/2 , which is characteristic of purely diffusive behavior.27 In marked
contrast, the leading asymptotic behavior of Pj (t) for a biased random walk is given
√
−1/2 −λt(1−2 pq) √
by Pj (t) ∼ t e . Since 1 > 2 pq when p
= q, this shows that Pj (t) decays
exponentially with time for all j.
dP (j, t) λ
= P (j + δ, t) − λP (j, t) ,
dt 2d δ
loss term
gain term
where δ stands for a nearest-neighbor vector of the site j. (That is, δ has one of its d
components equal to ±1, and all other components equal to 0.) The initial condition
is
1, if j = 0
P (j, 0) =
0, if j
= 0.
27
We’ll study the diffusion equation and its properties later on.
85
Define the generating function
f (z1 , . . . , zd , t) = P (j, t) z1j1 z2j2 · · · zdjd ,
j
∂f z + z −1 + · · · + z + z −1
1 d
=λ 1 d
− 1 f,
∂t 2d
with the initial condition f = 1 at t = 0.
(b) The solution to this equation is clearly an exponential that factors into a product
of generating functions of modifed Bessel functions. Show that
The leading asymptotic (t → ∞) behavior of P (j, t) follows from that of the modified
Bessel function. We find P (j, t) ∼ t−d/2 . This specific power-law decay of the proba-
bility is characteristic of purely diffusive motion.
13. The generalization of the solution above to include any directional bias is straight-
forward. Let pi and qi be the
probabilities of a jump in which ji changes to ji + 1 and
ji − 1, respectively. Then d1 (pi + qi ) = 1.
d
√
−λt
P (j, t) = e (pi /qi )j/2 Iji 2λt pi qi .
i=1
86
8 Fourier transforms
Fourier integrals: As you know, periodic functions can be expanded in Fourier
series. Recall that, if the fundamental interval of the function f (x) is (a , b) so that
the period is L = (b − a), the expansion and inversion formulas are
∞ b
f (x) = (1/L) fn e 2πnix/L
and fn = dx f (x) e−2πnix/L .
n=−∞ a
The function f(k) is obtained from the function f (x) by acting upon the latter with
an integral operator whose kernel is exp (−ikx). Similarly, the function f (x) is ob-
tained from the function f(k) by acting upon the latter with an integral operator whose
kernel is (2π)−1 exp (ikx). The functions f (x) and f(k) are Fourier transforms of
each other.28 A Fourier transform is an example of an integral transform. A func-
tion f (x) has a Fourier transform if it satisfies conditions analogous to the Dirichlet
conditions for periodic functions. Broadly speaking, if
(i) f (x) has at most a finite number of finite discontinuities or jumps, and
28
Which of the two we call the transform, and which we call the inverse transform, is a matter of
convention.
87
∞
(ii) f (x) is absolutely integrable in (−∞ , ∞), so that −∞
dx |f (x)| < ∞ (or f (x) ∈
L1 (−∞ , ∞)),
I repeat that these are just matters of convention,29 but it is quite important to choose
some specific convention and stick to it consistently, in order to avoid errors.
Parseval’s formula for Fourier transforms: If f (x) is not only integrable but is
also square-integrable, i.e., if f (x) ∈ L2 (−∞ , ∞), it follows on applying Parseval’s
formula that its Fourier transform f(k) is also an element of L2 (−∞ , ∞): we have
∞ ∞
1
dx |f (x)| =
2
dk |f(k)|2 .
−∞ 2π −∞
(This relation is just the continuum analog of the discrete version of Parseval’s formula
for Fourier series.) The connection between L2 functions and Fourier tranforms should
be familiar to you from elementary quantum mechanics. The Fourier transform, as
applied to elements of the function space L2 , may be regarded as a change of basis.
In the context of quantum mechanics, Parseval’s formula implies that, if the position-
space wave function of a particle is normalizable, then so is its momentum-space wave
function, because these two wave functions form a Fourier transform pair.
Fourier transform of the δ-function: The Dirichlet conditions and the conditions
stated above are sufficient conditions for a function f (x) to have a Fourier series ex-
pansion or a Fourier transform, as the case may be. They are by no means necessary
conditions. For instance, numerous functions that are integrable, but not absolutely
integrable (i.e., their absolute values are not integrable) have Fourier transforms. Func-
tions that are more singular than permitted by the Dirichlet conditions may also have
Fourier representations. The theory of generalized functions or distributions is, in fact,
very closely linked with the Fourier transforms of these objects. There is a highly-
developed area of mathematics, harmonic analysis, dealing with these matters and
their generalizations.
29
For instance, as you’ll see below, the kernels (2π)−1/2 eikx and (2π)−1/2 e−ikx are often used in
the expansion formula and the inversion formula, instead of the kernels (2π)−1 eikx and e−ikx that we
have used.
88
The Dirac δ-function, which is so useful in applications, also has a Fourier transform.
Since we know that ∞
1
δ (x) = dk eikx ,
2π −∞
it follows that the Fourier transform of δ (x) is just unity, i.e., δ (k) = 1.
1. Sketch the functions f (x) listed below, and show that their Fourier transforms are
as given. Here a and σ are positive constants, while b and µ are real constants. θ(x)
is the unit step function.
−4iabk
(a) f (x) = e−a|x| sin bx =⇒ f(k) = .
(a2 + b2 + k 2 )2 − 4b2 k 2
2a(a2 + b2 + k 2 )
(b) f (x) = e−a|x| cos bx =⇒ f(k) = .
(a2 + b2 + k 2 )2 − 4b2 k 2
[θ (x + a) − θ (x − a)] sin ka
(c) f (x) = =⇒ f(k) = .
2a ka
sin ax
(d) f (x) = =⇒ f(k) = π [θ (k + a) − θ (k − a)].
x
1
f(k) = e−iµk− 2 k σ .
2 /(2σ 2 ) 1 2 2
(e) f (x) = √ e−(x−µ) =⇒
2π σ 2
The functions in (c) and (e) above can be regarded as normalized probability density
functions corresponding to a random variable x. The first of these corresponds to
a uniform distribution, while the second corresponds, of course, to the normal or
Gaussian distribution. The Fourier transform of the probability density of a random
variable is called its characteristic function. Knowing the characteristic function of
a random variable is equivalent to knowing all its moments and cumulants.
2. Sketch the functions f(k) listed below, and show that their inverse Fourier trans-
forms f (x) are as given. (σ, k0 and λ are positive constants.)
1 1
(a) f(k) = e− 2 (k−k0 ) σ =⇒ f (x) = √
2 2 2 2
eik0 x e−x /(2σ ) .
2πσ 2
k i(1 − cos k0 x)
(b) f(k) = θ (k0 − |k|) =⇒ f (x) = .
|k| πx
1 − cos (k0 x) + k0 x sin (k0 x)
(c) f(k) = θ (k0 − |k|) (k0 − |k|) =⇒ f (x) = .
2πx2
λ
(d) f(k) = e−λ|k| =⇒ f (x) = .
π (x2 + λ2 )
89
The function f (x) in case (d) is a ‘Lorentzian’. It is is the probability distribution
function corresponding to the Cauchy distribution.
The convolution theorem for Fourier transforms is a basic and most useful rela-
tionship. Given two functions f (x) and g(x), the function
∞ ∞
h(x) = dx f (x ) g(x − x ) = dx f (x − x ) g(x )
−∞ −∞
is called the convolution of the functions f (x) and g(x). This relationship is sometimes
written as h = f ∗ g . (This should not be confused with ordinary multiplication of the
functions concerned.)
3. Establish the convolution theorem for Fourier transforms: If f (x) and g(x) have
Fourier transforms f(k) and
g (k) respectively, then the Fourier transform of their con-
volution h(x) is given by h(k) = f(k)
g (k).
Thus, the Fourier transform of the convolution of two functions is just the product of
the Fourier transforms of the two functions. The Fourier transform operation converts
90
convolution to an ordinary multiplication.
4. It should be clear from the foregoing that this works in reverse, too. Consider the
Fourier transform of the convolution of f and g, namely, of the function of k given by
∞ ∞
H(k) = dk f(k )
g (k − k ) ≡ dk f(k − k )
g (k ).
−∞ −∞
Check that the inverse Fourier transform of H(k) is simply the product H(x) =
f (x) g(x).
5. Generalization of Parseval’s formula: Let f (x) and F (x) be two good func-
tions30 of x, i.e., they have derivatives of all orders for all x, and vanish (along with
their derivatives of all orders) as x → ∞. Show that
∞ ∞
1
dx f (x) F (−x) = dk f(k) F(k) .
−∞ 2π −∞
Note that (i) neither side of this equation is in the form of a convolution of two func-
tions; (ii) there is a minus sign present in the argument of one of the functions in the
integrand; and (iii) each side of the equation is a number, and not a function. We’ll
use this result in the sequel, in the derivation of the Poisson summation formula.
Note that Parseval’s formula itself follows as a special case of the result above. All you
have to do is to identify F (−x) with f ∗ (x), and observe that
Iterates of the Fourier transform operator: We’ve seen that the Fourier trans-
form f(k) of a function f (x) may be regarded as the result of applying a certain integral
operator, namely, the Fourier transform operator F , to the function. The kernel of
the operator is just e−ikx . In order to make the notion precise, we need to specify a
function space, such that both the function f (x) and its transform f(k) belong to the
same space. As we know, the space L2 (−∞ , ∞) satisfies this requirement. Let us
therefore restrict ourselves to this function space in what follows.
30
I believe this simple but expressive terminology is due to M. J. Lighthill.
91
(Observe the notation in this equation! I have written x for the argument of the output
function after the Fourier transform operation is performed.) Something very interest-
ing happens when the operator F is iterated i.e., applied repeatedly to a function. We
have
∞
[F f ](x) ≡ [F F f ](x) = [F f ](x) =
2
dz e−ixz f(z)
−∞
∞ ∞ ∞ ∞
−ixz −izy
= dz e dy e f (y) = dy f (y) dz e−i(x+y)z
−∞
∞
−∞ −∞ −∞
= dy f (y) 2π δ (x + y) = 2π f (−x) .
−∞
Now, we know that the parity operator P changes the sign of the argument when
acting on any function of x. That is, (Pf )(x) ≡ f (−x) . The result above shows that
the square of the Fourier transform operator is just 2π times the parity operator. In
symbols,31
F 2 = 2π P.
Moreover, since the square of the parity operator is obviously just the unit operator I,
we have the operator relationship
F 4 = (2π)2 P 2 = (2π)2 I.
The Fourier transform operator is thus proportional to a ‘fourth root’ of the unit op-
erator! These results find applications in Fourier optics.
Recall that the Hamiltonian of the quantum mechanical linear harmonic oscillator
has exactly the same form in the position basis and in the momentum basis: In units
such that m, ω and are all equal to unity, the Hamiltonian is represented by the
operators
1 d2 1 d2
− 2 + x2 and − 2 + p2 ,
2 dx 2 dp
respectively, in the position basis and the momentum basis. As a result, the normalized
position-space wave functions Φn (x) and the momentum-space wave functions Φ n (p)
representing the eigenstates of this Hamiltonian must be identical in form. But we also
know that Φn (x) and Φ n (p) are elements of L2 (−∞ , ∞), and that they are Fourier
transform pairs. It follows that these are precisely the eigenfunctions of F that we’re
31
The factor of 2π is a consequence of the particular Fourier transform convention I have used. It
should be obvious that the factor can be made equal to unity by a suitable choice of convention.
92
looking for. It remains to prove this assertion directly, and also to find the eigenvalues
of F explicitly. Let’s now do so.
Multiply both sides of the last equation by exp (−ikx − 12 x2 ), and integrate over x
from −∞ to ∞. The left-hand side becomes a Gaussian integral that can be evaluated.
The result is
∞ ∞
tn % −x2 /2
∞
1/2 −k 2 /2 −2ikt+t2 −ikx −x2 /2
& tn
(2π) e e = dx e e Hn (x) = F e Hn (x) .
n=0 −∞
n! n=0 n!
But the second exponential factor on the left-hand side is again a generating function
for Hermite polynomials. We have
∞
(−it)n
−2ikt+t2 2k(−it)+(−it)2
e =e = Hn (k) .
n=0
n!
2
Hence the function
√ e−x /2 Hn (x), or this function multiplied by the normalization con-
stant (2n n! π )−1/2 , is an eigenfunction of the Fourier transform operator F in the
space L2 (−∞ , ∞). Here n runs over the values 0, 1, 2, . . . . The corresponding eigen-
value is (2π)1/2 (−i)n .
• The eigenvalues of the operator (2π)−1/2 F are (−i)n , where n = 0, 1, 2, . . . .
93
6. Work through the steps of this derivation.
Had we used the alternative (and more symmetrical) Fourier transform convention
∞ ∞
1 −1 1
[F f ](k) = √ dx e−ikx
f (x), [F f ](x) = √ dk eikx f(k),
2π −∞ 2π −∞
we would have found that F 4 = I, and the eigenvalues of F would have been simply ±1
and ±i. Indeed, the customary normalization of quantum mechanical wave functions
uses precisely this convention. But the convention chosen here is used in many other
applications of Fourier analysis, and so let’s stay with it.
The adjoint of an integral operator: Let f (x) be the function representing the
vector | f in the given function space (here, L2 (−∞ , ∞)). If
∞ K is the integral operator
concerned, the vector K | f is represented by the function −∞ dy K(x, y) f (y), where
K(x, y) is the kernel of the operator. The adjoint of K, denoted by K† , is of course to
be identified by applying the condition g | Kf = K† g | f for every pair of elements
94
| f , | g in the function space. But
∞ ∞
∗
g | Kf = dx g (x) dy K(x, y) f (y)
−∞ −∞
∞ ∞
= dy dx g ∗(x) K(x, y) f (y) (changing the order of integration)
−∞∞ −∞∞
= dx dy K(y, x) g ∗(y) f (x) (re-labeling x ↔ y)
−∞∞ −∞∞
= dx dy [K ∗ (y, x) g(y)]∗ f (x) = K† g | f .
−∞ −∞
Returning to the case at hand, we see that (2π)−1/2 F has the kernel K(k, x) =
(2π)−1/2 e−ikx , while its inverse [(2π)−1/2 F ]−1 has the kernel (2π)−1/2 eikx . But the
latter quantity is precisely K ∗ (x, k). We therefore have the operator identity
[(2π)−1/2 F ]−1 = [(2π)−1/2 F ]† .
In other words, the operator (2π)−1/2 F is a unitary operator. It is not surprising,
then, that all its eigenvalues lie on the unit circle in the complex plane, just as those
of a unitary matrix do!
Hence the Fourier transform in L2 (−∞ , ∞) is not only a change of basis in the
space, but also a unitary transformation. This fact has implications in quantum
mechanics. For instance, it guarantees that the description of a particle (or a system
of particles) in position space and in smomentum space are unitarily equivalent—i.e.,
you can use either description without altering the underlying physics.
The Fourier transform in d dimensions: The Fourier transform and most of the
results in the foregoing are generalized in a straightforward manner to functions of
r ∈ Rd , where d = 2, 3, . . . . We have
1
f (r) = d k e f (k) ⇐⇒ f(k) = dd r e−ik·r f (r).
d ik·r
(2π)d
Here dd k and dd r denote the volume elements in d-dimensional k-space and r-space,
respectively. The Fourier representation of the d-dimensional δ-function is
1
(d)
δ (r) = dd k ei k·r .
(2π)d
32
Observe how reminiscent this is of the situation in the case of matrices. If the (ij)th element of
a matrix A is aij , the (ij)th element of its hermitian adjoint A† is a∗ji . The close analogy between
integral operators and matrices (as operators in an LVS) plays a crucial role in the theory of integral
equations.
95
Fourier expansions of vector-valued functions can also be written down in an anal-
ogous manner. Thus, if u(r) is a vector field in Rd , we have
1
u(r) = d
d ke uik·r
(k) ⇐⇒ u (k) = dd r e−ik·r u(r) .
(2π)d
One of the great advantages of the Fourier expansion of functions now becomes evident.
Consider the usual case, d = 3. A vector field is generally specified by differential
equations for its divergence and curl, respectively.33 We have
1
∇ · u(r) = 3
(k)
d3 k eik·r ik · u
(2π)
and
1
∇ × u(r) = (k).
d3 k eik·r ik × u
(2π)3
' (
Further, the set of functions eik·r | k ∈ R3 forms an orthonormal basis for integrable
functions of r ∈ R3 . Hence partial differential equations for the divergence and curl of
a vector field u(r) reduce to algebraic equations for its Fourier transform u (k). The
utility of this result should now be obvious to you.
The Poisson summation formula is a very useful result that has many applications.
For instance, it can be used to sum many infinite series. It is also related to much deeper
mathematical results (which we do not go into here).
Consider the infinite periodic array of δ-functions (a Dirac comb) given by
∞
F (x) = δ (x − nL),
n=−∞
96
But
1 x
δ n−
δ (x − nL) = δ (nL − x) = .
L L
We thus have a useful identity relating a sum of equally-spaced δ-functions to a sum
of exponentials, namely,
x
∞ ∞
2πnix
δ n− = exp .
n=−∞
L n=−∞
L
It is worth pausing once again for a moment to think about this most remarkable for-
mula. Note that neither side of this equation is anything like an absolutely convergent
series!
But, as we have seen already, an infinite sum of exponentials of this kind can be written
as an infinite sum of δ-functions. Using this relationship, we get
∞
kL
F (k) = δ n− .
n=−∞
2π
The final step is to insert the expressions for F (x) and F(k) in the identity34
∞ ∞
1
dx f (x) F (−x) = dk f(k) F(k).
−∞ 2π −∞
Using the fact that the function F (x) in the present instance is an even function of x,
we get
∞ ∞ ∞ ∞
1 kL
dx f (x) δ (x − nL) = dk f (k) δ n− ,
−∞ n=−∞
2π −∞ n=−∞
2π
where f (x) is any arbitrary function with a Fourier transform f(k). Finally, therefore,
∞
∞
f (nL) = (1/L) f(2πn/L) .
n=−∞ n=−∞
34
The identity was derived for good functions, but we’re using it here for singular functions like the
δ-function. I’ll merely state here without proof that this step can be justified.
97
Here L is an arbitrary positive parameter. This is the famous Poisson summation
formula It is helpful in the summation of series in cases when the left-hand side is
difficult to evaluate, but the right-hand side is more tractable; or vice versa. Note
that the parameter L occurs in the numerator of the argument of the function on the
left-hand side, but on the right-hand side it occurs in the denominator. Hence one of
the two representations may be useful for small values of L, while the other is useful
for large values of L, in physical applications.
Some illustrative examples: Here are some instances of the use of Poisson’s sum-
mation formula to derive useful identities and to sum certain infinite series.
This identity, due to Jacobi, is so useful and important that it is sometimes called the
Poisson summation formula itself!35 For example, in the context of the phenomenon of
diffusion, the parameter λ2 is proportional to Dt, where D is the diffusion coefficient
and t is the time. The Poisson summation formula can then be used to obtain valuable
insight into the behavior of the solution to the diffusion equation for both small t and
large t.
98
Setting L = 1 in the Poisson summation formula, we therefore have
∞
cos (nb) ∞
π % −a |2πn−b| &
−a |2πn+b|
2 + a2
= e + e .
n=−∞
n n=−∞
2a
Setting b = 0 in this relation, we recover an important and useful result that we have
already derived using contour integration, namely,
∞
1 π 1
S(a) ≡ 2 2
= coth πa − .
n=1
n +a 2a πa
As usual, a relation of this sort may be regarded as a relation between two analytic
functions—in this case, of the variable a. Recall that the quantity in curly brackets
on the right-hand side is the Langevin function. It occurs, for instance, in the el-
ementary
∞ theory of paramagnetism. Passing to the limit a → 0 once again gives us
n=1 1/n 2
≡ ζ(2) = π 2 /6. Repeatedly differentiating S(a) with respect to a and pass-
ing to the limit a → 0 enables us to find ζ(4) = π 4 /90, ζ(6) = π 6 /945, . . . .
Finally, it turns out that the generalizations of the Poisson summation formula
are related to deep results in advanced mathematics. I merely mention some of these
results, to whet your curiosity: the asymptotic behavior of the heat kernel, the spectra
of the Laplacian operator and general elliptic operators on manifolds, Eisenstein series,
automorphic functions, the Selberg trace formula, and so on.
99
9 QUIZ 1
1. Are the statements in quotation marks true or false?
(b) Let u and v denote the real and imaginary parts of an analytic function of
z = x + iy.
“The curves u(x, y) = constant and v(x, y) = constant intersect each other
at right angles.”
∞
(e) “The radius of convergence of the power series n=1 n1/n z n is zero.”
(f) “The function sin (π/z) has an accumulation point of poles at z = 0.”
(g) “The relation Γ(z) Γ(1 − z) = π cosec πz is only valid in the region 0 <
Re z < 1.”
(i) “The power series ∞ n 4
n=1 z /n is absolutely convergent at all points inside
and on the unit circle |z| = 1.”
∞
(j) “The series (n + 1)z−1 converges in the region Re z > 0 .”
n=0
√ √
(k) Consider the Möbius transfomation z → w = (2z + 3 )/( 3 z + 2).
“This is a hyperbolic Möbius transformation.”
√ √
(l) Consider z → w = (2z + 3 )/( 3 z + 2) once again.
“The transformation maps the circle |z| = 1 to the circle |w| = 1.”
100
(m) Legendre’s differential equation is (1 − z 2 )φ − 2zφ + ν(ν + 1)φ = 0.
“Since this equation is invariant under the interchange ν ↔ −ν − 1, all its
solutions must also be invariant under this interchange.”
3/2
(n) “The function f (t) = et , where t ≥ 0, has no Laplace transform.”
∞ ∞
(o) “If [Lf ](s) = 0
dt e−st f (t), then [L2 f ](s) = 0
dt f (t)/(s + t).”
(p) “The product Γ(z) ζ(z) tends to a finite, nonzero limit as z → −2n, where
n = 1, 2, . . . .”
its Laplace
(q) Let φ(t) be a linear, causal, retarded response function, and φ(s)
transform.
“The corresponding dynamic susceptibility χ(ω) is the analytic continuation
to s = −iω.”
of φ(s)
(r) “The logarithmic derivative of the Riemann zeta function, ζ (z)/ζ(z), has a
simple pole at z = 1 with residue equal to −1.”
(s) “The only pole of the logarithmic derivative of the Riemann zeta function
is at z = 1.”
(a) The real part of an entire function f (z) is given by u(x, y) = (cosh x) (cos y).
Hence the function is f (z) = · · · .
101
(d) The residue at z = ∞ of the polynomial p(z) = z n + a1 z n−1 + · · · + an is
Res p(z) = · · · .
z=∞
(e) Let
C denote the circle |z| = 2 traversed once in the positive sense. Then
C
dz/(z 4
− 1) = ···.
(f) Let a and b be two different complex numbers, each with nonzero real and
imaginary parts. The radius of convergence of the power series
∞
Γ(n + a) Γ(b) z n
f (z) =
n=0
Γ(a) Γ(n + b) n!
is R = · · · .
∞
∞
(g) Given that 1/n4 = π 4 /90, it follows that 1/(2n + 1)4 = · · · .
n=1 n=0
∞
∞ that 0 dx (sin
(h) Given kx)/x = 12 π (where k > 0), the value of the integral
0
dx (1 − cos x)/x = · · · . (Hint: Integrate k over a suitable range.)
2
1
(j) The value of the integral 0
dt t−1/2 (1 − t)−1/2 = · · · .
(k) Let an arbitrary initial point z (0) in the complex plane be mapped to the
point z (n) under n iterations of the Möbius transfomation
√ √
z → (2z + 3 )/( 3 z + 2).
(l) Previous question continued: The exceptional point that does not tend to
the limit point above is z = · · · .
(m) Under the Möbius transformation z → w = (2z + 3)/(z + 2), the circle
|z + 2| = 1 is mapped to the circle · · · .
102
√
(n) The function f (z) = z ln [(z − 1)/(z + 1)] has branch points at z = · · · .
(q) Given that the Laplace transform of sin t is 1/(s2 + 1), it follows that the
Laplace transform of sinh t is · · · .
∞
tn
2tz−t2
e = Hn (z) .
n=0
n!
(s) Let
1, |x| ≤ 1
f (x) =
0 |x| > 1.
If f(k) denotes the Fourier transform of f (x), the value of the integral
∞
−∞
dk |f(k)|2 = · · · .
(t) Consider a random walk on an infinite linear lattice whose sites are labelled
by the integers. The walker jumps from any site j to j −1 with a probability
per unit time given by λq, and from j to j + 1 with a probability per unit
time given by λp; further, the walker stays at the site j with probability
per unit time given by λr. Here p, q and r are positive constants satisfying
p + q + r = 1, and λ is a positive constant with the physical dimensions of
(time)−1 . Let P (j, t) be the probability that the walker is at the site j at
time t. The differential equation satisfied by P (j, t) is dP (j, t)/dt = · · · .
103
Quiz 1: Solutions
1. (a) T
(b) T
(c) T
(d) F
(e) F
(f) F
(g) F
(h) T
(i) T
(j) F
(k) T
(l) T
(m) F
(n) T
(o) T
(p) T
(q) T
(r) T
(s) F
(t) T
104
(k) 1
(l) −1
(m) |w − 2| = 1
(n) −1, 0, 1 and ∞
(o) 1, i, −1, −i and ∞
(p) I1 (2)
(q) 1/(s2 − 1)
dn n
2tz−t2 n z2 d 2
(r) Hn (z) = n
e , which simplifies to H n (z) = (−1) e n
e−z .
dt t=0 dz
(s) 4π
dP (j, t)
(t) = λ p P (j − 1, t) + q P (j + 1, t) − (p + q) P (j, t)
dt
105
10 The fundamental Green function for ∇2
The Laplacian operator ∇2 appears in all the standard and important linear partial dif-
ferential equations of elementary mathematical physics: Laplace’s equation, Poisson’s
equation, Helmholtz’s equation, the diffusion or heat equation, and the wave equation.
We’ll now consider Poisson’s equation, which requires the determination of the Green
function of the Laplacian operator. Let’s start with a brief recapitulation of the essen-
tials of the Green function method.
D | f = | g .
where the ci are constants, and the kets |hi are the linearly independent solutions of
the homogeneous equation
D |hi = 0 or Dx hi (x) = 0.
In function space, the inverse of the operator D is represented by the inverse of the
differential operator Dx . In general, the latter is an integral operator. To see this,
take the scalar product of both sides of the formal solution for | f above with x |.
We have
x | f ≡ f (x) = x |D−1 | g + ci x |hi
i
b
= dx x |D−1| x x | g + ci hi (x)
a i
b
= dx G(x, x ) g(x ) + ci hi (x),
a i
106
where
G(x, x ) ≡ x |D−1|x .
G(x, x ) is the Green function of the differential operator Dx . It is just the ‘matrix
element’ of the operator D−1 between the states x | and | x , and is called the kernel
of the integral operator D−1 .
The fact that G(x, x ) represents (in function space) the inverse of the abstract
operator D−1 means that it, too, satisfies the same differential equation as f (x), but
with a δ-function as the inhomogeneous term. To see this, we start with
But
b
−1
x |D D | x = x |D dy | y y | D−1 | x
a
b
= dy x |D| y G(y, x ) = dy Dx x| y G(y, x )
a
= dy Dx δ(x − y)G(y, x ) = Dx G(x, x ).
Dx G(x, x ) = δ(x − x ).
In order to write down the solution for f (x) explicitly, we need to find the Green
function G(x, x ). This must be done by solving the differential equation for G, using
appropriate boundary conditions and adjusting the values of the constants ci . In this
manner, we arrive at the unique solution that satisfies the given boundary conditions
on f (x).
1. Here’s a simple example that shows how the Green function method works. Con-
sider the ordinary differential equation
d2
f (x) = g(x),
dx2
107
where x ∈ [0, 1], and g(x) is a given function of x. It is required to find the solution
f (x) that satisfies the general linear boundary conditions
In essence, this is the Green function method. All that has been said above can be
generalized to linear ordinary differential equations in a complex variable z, and fur-
ther, to the case of linear partial differential equations in several variables—e.g., when x
is replaced by r in any number d of spatial dimensions, or by (r, t) in (d+1) dimensions.
Poisson’s equation; the fundamental Green function for ∇2 : Let’s turn now to
a problem of great physical importance: the determination of the Green function of the
Laplacian operator, required for the solution of Poisson’s equation. In effect, we seek
the inverse of the ∇2 operator. To start with, consider the standard three-dimensional
case. Subsequently, we’ll go on to the cases d = 2 and d ≥ 4.
As you know, the electrostatic potential φ(r) in the presence of a given static charge
density ρ(r) satisfies Poisson’s equation, ∇2 φ(r) = −ρ(r)/0 . The solution satisfying
natural boundary conditions,
3 namely, φ(r) = 0 as r → ∞ along any direction in space,
−1
is φ(r) = (4π0 ) d r ρ(r )/|r − r |. This is just Coulomb’s Law together with the
Consider the general form of Poisson’s equation together with natural boundary
conditions, namely,
g(r) is a given function that acts as the ‘source’ term for the scalar field f (r). We’re
interested here in the PI, given by
f (r) = d3 r G(r, r ) g(r ),
108
where the Green function G satisfies the differential equation
I’ve used a subscript on the gradient operator in order to indicate the variable with
respect to which the differentiation is to be performed. We require the solution that
vanishes as r → ∞. Note that G may be interpreted (apart from a constant of
proportionality) as the Coulomb potential due to a point charge at the source point
r . Observe that
(i) the operator ∇2 is translationally invariant—shifting r by r does not change the
operator;
(ii) the δ-function on the right-hand side is a function of the difference r − r ; and
finally,
G(k) = −1/k 2 .
109
∞
vector R. You will need the Dirichlet integral 0 dk (sin kR)/k = 12 π sgn R = 12 π. The
final result is
1 1
G(r − r ) = − =− .
4πR 4π|r − r |
Note that the Green function is actually a function of R alone. The PI in the solution
of Poisson’s equation is then
1 g(r )
f (r) = − d3 r .
4π |r − r |
where rs = min (r, r ) and rg = max (r, r ). Insert this expansion in the solution for
f (r), and and interchange the order of summation and integration. The integrations
over the angular variables θ and ϕ can then be carried out immediately by using the
fact that √
dΩ Ylm
∗
(θ , ϕ ) = 4π δl,0 δm,0 .
Because of the Kronecker deltas, the dependence of f (r) on the angles θ and ϕ also
disappears, so that f (r) = f (r). Show that the solution simplifies to
∞ 2 ∞
r g(r ) 1 r 2
f (r) = − dr =− dr r g(r ) − dr r g(r ).
0 r g r 0 r
4. Check that the solution above satisfies Poisson’s equation, ∇2 f = g. You will need
the Laplacian in spherical polar coordinates, and also the formula for differentiation
under the integral sign.
110
the surface, apart from a multiplicative constant (= 1/0 , in SI units). A counterpart
obviously exists in gravitation as well. The inverse-square central force is the only
force law in three-dimensional space for which such a property holds good. Interest-
ingly enough, in spaces of other dimensionalities (d = 2, 4, . . .) too, there exist force
laws with an analogous property. They arise from the counterparts of the Coulomb
potential in those spaces. The case d = 2 is somewhat exceptional, as you will see, and
will be dealt with after we discuss the case d ≥ 3.
What is the analog of the Coulomb potential (or the inverse-square central force) in
a space of an arbitrary number of dimensions? The connection that I have just pointed
out leads to the consistent way to define such a potential, via Poisson’s equation for
the potential due to a point source (or charge):
The superscript in G(d) is to remind us that we are concerned with the Green function
in d-dimensional space. We’ll find the solution in two different ways: first, by using
Fourier transforms as in the d = 3 case derived above, and evaluating the resulting
integral the ‘hard’ way; second, by using a much simpler argument that will get us to
the same answer. This exercise is worth it because it is instructive.
Fourier-transforming the equation for G(d) (R) gives G(d) (k) = −1/k 2 , as before.
Remember that we are now working in d dimensions. Therefore
d
1 d k ik·R
G (R) = −
(d)
d
e .
(2π) k2
That is, the Coulomb potential in d dimensions, i.e., the fundamental Green function
of the Laplacian, is essentially the inverse Fourier transform of −1/k 2 .
5. The next task is to evaluate the d-dimensional above for G(d) (R). Once again,
This is most conveniently done in terms of the analog of spherical polar coordinates
in d dimensions, in k-space. These ultraspherical coordinates consist of (i) the
magnitude |k| = k, (ii) a set of (d − 2) ‘polar’ angles θ1 , θ2 , . . . , θd−2 , and (iii) an
‘azimuthal’ angle ϕ. The ranges of these variables are given by
111
The Cartesian components k1 , k2 , . . . , kd are related to the polar coordinates by
k1 = k cos θ1
k2 = k sin θ1 cos θ2
k3 = k sin θ1 sin θ2 cos θ3
··· = ························
kd−1 = k sin θ1 sin θ2 . . . sin θd−2 cos ϕ
kd = k sin θ1 sin θ2 . . . sin θd−2 sin ϕ.
We can choose the orientation of the axes such that R is along the direction of the
first coordinate k1 , so that k · R = kR cos θ1 . The integration over ϕ gives a factor
of 2π. Carry out the integration over each of the ‘polar’ angles θ2 , . . . , θd−2 using the
integral √
π
π Γ 12 (r + 1)
r
dθ sin θ = .
0 Γ 1 + 12 r
But the integral over the last angular coordinate, θ1 , is more complicated: It turns out
to be essentially a representation of the Bessel functionof the first kind. The formula
you require is
1 ν π
z
Jν (z) = √ 2 dθ e±iz cos θ sin2ν θ, Re ν > − 12 .
π Γ ν + 12 0
Insert these results and simplify the expression obtained, to arrive at the following
expression: ∞
1
G (R) = −
(d)
d/2 (d−2)/2
dk k (d−4)/2 J d −1 (kR).
(2π) R 0
2
Thus G(d) (R) has been reduced to a single integral. It is obvious from the last equation
that the Green function is only a function of the magnitude R of the vector R. Let’s
therefore write it as G(d) (R) from now on.
A divergence problem: A difficulty crops up now. The integral in the last equation
above does not converge for arbitrary values of the dimensionality
∞ d. As you have
seen, the case d = 3 already involves the Dirichlet integral 0 dk (sin kR)/k, which
112
is convergent, but not absolutely convergent. It is finite only because the sine func-
tion oscillates in sign, and there is some cancellation between positive and negative
contributions
∞ to the integral. If the integrand is replaced by its magnitude, we find
that 0 dk | sin kR|/k diverges logarithmically owing to the slow (∼ k −1 ) decay of the
integrand as k → ∞. For d > 3, you may expect the divergence of the relevant integral
to get worse, essentially because the volume element dd k in d dimensions involves a
factor k d−1 . This factor increases rapidly with k for larger values of d.
Dimensional regularization: Consider, first, the conditions under which the integral
∞
dk k (d−4)/2 J d −1 (kR)
2
0
36
The actual asymptotic form is [2/(π kR)]1/2 times a cosine function whose magnitude does not
exceed unity.
113
At the lower limit of integration, the situation is different. We know that
dk k r < ∞ provided Re r > −1.
0
Jν (kR) ∼ (kR)ν .
The integrand therefore behaves like k d−3 in the neighborhood of k = 0, and the
integral converges if
1
2
(Re d − 4) + 12 Re d − 1 > −1, that is, if Re d > 2.
Thus, the integral representing G(d) (R) converges in the range 2 < Re d < 3 of
the parameter d. For other values of d, the integral is infinite. Let’s now see how
dimensional regularization helps us extract sensible results from the formally divergent
integral we are faced with. The underlying idea is as follows.
(i) The dimension d itself is treated as a complex variable. In the region of the
complex d-plane in which the integral converges, it defines an analytic function
of d.
(ii) Analytic continuation is then used to go outside this region in the d-plane.
(iii) There are likely to be singularities (specifically, poles) present at the boundaries
of the original region of convergence. If such a singularity occurs at a physical
value of d, the regular part of the function (i.e., the function with the singular part
subtracted out) is supposed to represent the value of the function at the point
concerned.
6. First of all, keeping d in the region 2 < Re d < 3, we must evaluate the integral
in the formula for G(d) (R). It turns out to be expressible in closed form, on using the
following formula:
∞
2µ Γ 12 (ν + 1 + µ)
dx x Jν (ax) = µ+1 1
µ
, Re µ < − 12 , Re (µ + ν) > −1
0 a Γ 2 (ν + 1 − µ)
114
where a is a positive constant. First check that, when applied to the integral at
hand, the conditions on the parameters µ and ν amount to precisely the restriction
2 < Re d < 3. Next, use the formula above to show that
Γ 12 d − 1
G (R) = − d/2 d−2 .
(d)
4π R
But this expression for G(d) (R) is now in an explicit closed form that is analytic even
to the right of Re d = 3. Setting d = 3, it trivially checked that G(3) (R) = −1/(4πR),
as we have found already. You can now proceed to set d = 4, 5, . . . in the final result
to write down the fundamental Green function of the Laplacian in these dimensions.
We conclude that:
• the fundamental Green function of the Laplacian (i.e., the Coulomb potential)
in d ≥ 3 spatial dimensions is proportional to 1/Rd−2 .
37
This theorem is not restricted to three-dimensional space! It is valid in any number of dimensions,
as it is really a form of the fundamental theorem of the calculus.
115
But the field required is spherically symmetric, and has only a radial component FR
which, moreover, depends only on the magnitude R. Hence
F · dS = FR dS = Sd (R) FR = 1,
8. Determine Sd (R). You can do so, for instance, by integrating the ‘solid angle’
element
def.
dΩ(d) = (sind−2 θ1 ) (sind−3 θ2 ) . . . (sin θd−2 ) dθ1 . . . dθd−2 dϕ,
over the full ranges of the angular variables, and multiplying the result by Rd−1 . The
answer is
2π d/2 Rd−1
Sd (R) = .
Γ 12 d
This is precisely the expression found earlier. As already pointed out, it is singular at
d = 2. The singularity is a simple pole.
The reason why an inverse-square central force (in three dimensions) leads to an
integral theorem like Gauss’ Law is also obvious now. The surface area of a sphere
increases with its radius like R2 . If the field drops off like 1/R2 , there is obviously an
exact compensation in the flux of the field across a sphere centered at the origin, and
the total flux becomes independent of the radius R. Precisely the same thing happens
in d dimensions, provided the field drops off like 1/Rd−1 , because the ‘surface’ of a hy-
persphere increases like Rd−1 . The potential must then decrease like 1/Rd−2 , exactly
116
as we have deduced.
(i) Write G(d) (R), which is an analytic function of d, in the form of a Laurent series
about the point d = 2, i.e.,
residue
∞
(d)
G (R) = + cn (d − 2)n .
(d − 2)
n=0
singular part regular part
(ii) Subtract out the singular part. Setting d = 2 in the regular part leaves behind
just the coefficient c0 , which is guaranteed to be the Green function G(2) (R) that we
seek. To find it, expand each of the d-dependent factors in the expression for G(d) (R),
except the pole factor (d − 2)−1 , in a Taylor series about d = 2. Retain only terms up
to the first order in (d − 2), to obtain
1 1 1
G(d) (R) = − + ln R + (ln π + γ) + O(d − 2),
2π(d − 2) 2π 4π
117
A direct derivation, once again: Should we believe the result obtained above, as
it seems to have been derived using a prescription that appears to be arbitrary? The
answer is ‘Yes’. Corroboration comes from the same direct physical argument as was
given for the case d ≥ 3. As in that case, we go back to the differential equation
Regard this as Poisson’s equation for the electrostatic potential due to a unit charge
(in suitable units) at R = 0. Let F ≡ ∇G(2) denote the planar vector field due to
this charge. Integrate both sides of Poisson’s equation over a circle of radius R. The
right-hand side of the equation becomes unity, of course. Gauss’ Theorem, applied to
the left-hand side, gives (using the circular symmetry of the field)
But FR = dG(2) /dR. Integrating with respect to R gives precisely the expression found
earlier, namely, G(2) (R) = (1/2π) ln R + constant.
• The fundamental Green function of the Laplacian (i.e., the Coulomb potential)
in 2-dimensional Euclidean space is proportional to the logarithm of R.
This seemingly simple fact has profound consequences in diverse areas of physics, such
as condensed matter physics and quantum field theory, among others. It even seems
to have a bearing on the phenomenon of quark confinement!
118
11 The diffusion equation
Fick’s laws of diffusion: Diffusion is the process by which an uneven concentration
of a substance gets gradually smoothed out spontaneously—e.g., a concentration of a
chemical species (like a drop of ink or dye) in a beaker of water spreads out ‘by itself’,
even in the absence of stirring. The microscopic mechanism of diffusion involves a very
large number of collisions of the dye molecules with those of the fluid, which cause the
dye molecules to move essentially randomly and disperse throughout the medium, even
without any stirring of the fluid. A macroscopic description of the process is based
on Fick’s Laws, and leads to a fundamental partial differential equation, the diffusion
equation. This equation serves as a basic model of phenomena that exhibit dissipa-
tion, a consequence of the irreversibility of macroscopic systems in time.
The local, instantaneous concentration ρ(r, t) of dye molecules satisfies the equation
of continuity, which is called Fick’s first law in this context:
∂ρ
+ ∇ · j = 0, (Fick’s I Law)
∂t
where j the ‘diffusion current density’. The crucial physical input is the specification of
this quantity. We assume that j is proportional to the local difference in concentrations,
i.e., to the gradient of the concentration itself. Thus
The positive constant D called the diffusion coefficient. It has the physical dimen-
sions of (length)2 /time. The minus sign on the right-hand side signifies the fact that
the diffusion occurs from a region of higher concentration to a region of lower con-
centration: That is, the diffusion current tends to make the concentration uniform.
Eliminating j, we get the diffusion equation for the concentration ρ(r, t):
∂
ρ(r, t) = D ∇2 ρ(r, t).
∂t
This equation is of first order in the time variable, and second order in the spatial
variables. It is a parabolic equation in the standard classification of second-order
partial differential equations. In order to find a unique solution to it, you need an ini-
tial condition that specifies the initial concentration profile ρ(r, 0), as well as boundary
conditions that specify ρ(r, t), for all t ≥ 0, at the boundaries of the region in which the
diffusion is taking place. The presence of the first-order time derivative in the diffusion
equation implies that the equation is not invariant under the time reversal transfor-
mation t → −t. Irreversibility is thus built into the description of the phenomenon.
119
the generalized susceptibility. In the present case, linear response implies that the
diffusion current that is set up in the medium as a result of the unequal concentrations
at different points is proportional to the gradient of ρ, rather than the gradient of some
nonlinear function of ρ (such as a power of ρ other than unity). Another example of
linear response, leading to an exact analog of the diffusion equation, is provided by
the phenomenological description of heat conduction. Given the initial temperature
distribution T (r, 0) of a body, the problem is to find the temperature distribution
T (r, t) at any later time t > 0. Analogous to Fick’s second law, it is assumed that
the heat flux is proportional to the negative of the temperature gradient (a linear
response). The constant of proportionality in this case is the thermal conductivity
of the body, κ. The equation for T (r, t) reads
∂
T (r, t) = κ ∇2 T (r, t).
∂t
This is why the diffusion equation is also known as the heat equation.
where δ (d) (r) is the d-dimensional δ-function. This means that the diffusing particle
starts at the origin at t = 0. In the context of the diffusion equation for the concen-
tration ρ(r, t), such an initial condition represents a point source of unit concentration
at the origin. In a space of infinite extent, we may take the starting point to be the
origin of coordinates without any loss of generality. The solution thus obtained is
the fundamental solution (or Green function) of the diffusion equation. As you’ll see,
it can be used to write down the solution corresponding to an arbitrary initial PDF
p(r, 0) = pinit (r) (or an initial concentration profile ρinit (r); all the results for p(r, t)
that follow are applicable, as they stand, to the case of ρ(r, t).)
120
r is concerned, it is natural to use Fourier transforms. To avoid confusion, let’s adopt
the following notation:
(You can tell which function we’re dealing with by looking at the arguments of the
functions concerned.) Taking the Laplace transform of both sides of the diffusion
equation, we get
s p(r, s) − p (r, 0) = D∇2 p(r, s).
Therefore
(s − D∇2 ) p(r, s) = δ (d) (r).
Now expand p(r, s) in a Fourier integral with respect to the spatial variable r, according
to
1 s).
p(r, s) = dd k eik·r φ(k,
(2π)d
The δ-function has the familiar Fourier representation
(d) 1
δ (r) = dd k eik·r .
(2π)d
Use these expressions in the equation for p(r, s) and equate the coefficients of the basis
vector eik·r in the space of functions of r. We must then have, for each k,
s) = 1, s) = 1
(s + Dk 2 ) φ(k, or φ(k, .
s + Dk 2
We thus obtain a very simple expression for the double transform φ(k, s). The trans-
forms must be inverted to find the PDF p(r, t). It is easier to invert the Laplace
transform first, and then the Fourier transform.38 The Laplace transform is trivially
inverted: recall that L−1 [1/(s + a)] = e−at . Therefore
−Dk 2 t 1 2
φ(k, t) = e , and hence p (r, t) = d
dd k eik·r e−Dk t .
(2π)
38
You can also invert the Fourier transform first, and then the Laplace transform. This procedure
leads to the same result, as it ought to, but it is needlessly complicated.
121
This d-dimensional integral factors into a product of d integrals upon writing r and k
in Cartesian coordinates. Each of the factors is the familiar ‘shifted’ Gaussian integral.
Show that the final result is
1 2
p (r, t) = d/2
e−r /(4Dt) .
(4πDt)
This is the fundamental Gaussian solution to the diffusion equation in d spatial dimen-
sions.
Solution for an arbitrary initial distribution: We can now write down the par-
ticular integral that solves the diffusion equation for any specified initial probability
density function. As a trivial generalization, we could take the initial instant to be any
t . Then, given the PDF pinit (r, t ) we have, for all t > t ,
1 d (r − r )2
p(r, t) = d r exp − pinit (r , t ).
[4πD(t − t )]
d/2 4D(t − t )
Thus, the PDF at any time t > t is an integral transform of the initial PDF. You will
recognize that the kernel of the transform is just the Green function of the operator
(∂/∂t − D∇2 ), and is given by
1 (r − r )2
G(r, t ; r , t ) = exp −
[4πD(t − t )]d/2 4D(t − t )
As I have mentioned already, the heat conduction equation has exactly the same form
as the diffusion equation. For this reason, the Gaussian kernel above is called the heat
kernel in the mathematical literature.
(Integration over the angular coordinates of the vector r yields zero.) The mean dis-
tance travelled by the particle does not vanish, of course, for any t > 0. In fact, all
122
the moments of the distance travelled in a given time interval t can be written down
easily. We find
∞
r (t) = d r r p(r, t) =
l d l
dr dΩd r d+l−1 p(r, t),
0
The noteworthy point is that the moments r l (t) increase with time like tl/2 . Moreover,
this is independent of the dimensionality d of the space in which the diffusion takes
place! In particular, the variance of the displacement is obtained by setting l = 2, and
is given by
Var (r) = r2 − r2 = r 2 = 2dDt.
In order to see how diffusion arises as the continuum limit of a random walk,
consider a random walk on a linear lattice. A little more generality is achieved by
considering a biased random walk, with different probabilities for jumps to the right
and left, respectively. on the lattice. This leads to a diffusion equation that is appro-
priate an important physical situation: namely, diffusion on a line in the presence of a
constant external force field. The force-free case is a special case of this equation. The
random walker starts at the site j = 0 at time n = 0. The probability of a step to the
right is α, while that of a step to the left is β = 1 − α. The probability that the walker
is at the site j (that is, at the point ja on the line) at time step n (that is, at time nτ )
is
P (ja, nτ ) = α P (ja − a, nτ − τ ) + β P (ja + a, nτ − τ ).
I have explicitly introduced the lattice constant a and the time step τ , so that it
becomes easy to see how the continuum limit (in both space and time) arises. This
happens when the quantities a, τ and α − β tend to zero simultaneously, as follows.
You will find it instructive to work through the steps outlined below.
123
Subtract P (ja, nτ − τ ) from each side of the equation for P (ja, nτ ). On the
right-hand side, re-write β as α − (α − β), and the coefficient of P (ja, nτ − τ ) as
−1 = −2α + (α − β). Collect terms suitably to get
P (ja, nτ ) − P (ja, nτ − τ )
= α P (ja − a, nτ − τ ) − 2P (ja, nτ − τ ) + P (ja + a, nτ − τ )
− (α − β) P (ja + a, nτ − τ ) − P (ja, nτ − τ ) .
Divide both sides by τ . Multiply and divide the first term on the right-hand side by
a2 , and the second term by a. Now let a → 0, τ → 0 and α − β → 0 (that is, let
α → 12 , β → 12 ), such that
a2 α a(α − β)
lim =D and lim = c,
τ τ
where D and c are finite, nonzero constants. Since α → 12 , the constant D is essentially
lim a2 /(2τ ). D has the physical dimensions of (length)2 /(time), while c has the
a,τ →0
physical dimensions of (length)/(time), i.e., a velocity. Further letting j → ∞ and
n → ∞ such that ja and nτ tend to the continuous variables x and t respectively, the
difference equation above for the probability P (ja, nτ ) reduces to the following partial
differential equation for the probability density function (PDF) p(x, t) of the position:
∂p (x, t)
j(x, t) = c p(x, t) − D .
∂x
The ratio D/c is a natural length scale in the problem, with the following physical
significance: it is a measure of the relative importance of the drift and diffusion contri-
butions to the motion of the colloidal particles. When the ratio c/D is multiplied by
124
a characteristic length L (such as that used in the definition of the Reynold’s number
in fluid flow), we get a dimensionless number Lc/D, called the Péclet number: it
quantifies the relative strengths of the advective transport rate and the diffusive
transport rate in fluid flow.
On the other hand, if the substance cannot leak out of the ends and stays confined in
the region b1 ≤ x ≤ b2 at all times, the diffusion current must vanish at the end-points.
Hence we must impose the so-called reflecting boundary conditions, for which I’ll
use the abbreviation RBC. The corresponding PDF will be written as pref (x, t). These
conditions imply that the flux or current, rather than the PDF itself, vanishes at the
boundaries. In the case of free diffusion, this means that
∂pref
= 0 at x = b1 and x = b2 , for all t ≥ 0. (RBC)
∂x
RBCs are a little more involved when both diffusion and drift are present. Recall that
the Smoluchowski equation for the PDF p(x, t) in the case of diffusion with drift in
one dimension is given by
∂p ∂j ∂p (x, t)
+ = 0, where j(x, t) = c p(x, t) − D .
∂t ∂x ∂x
125
Hence RBCs in this case are given by
∂pref
cpref − D = 0 at x = b1 and x = b2 . (RBC)
∂x
Other boundary conditions, or combinations of these, may be imposed, depending on
the physical problem.
and 2
1 −x /(4Dt) −(2b−x)2 /(4Dt)
pabs (x, t) = e − e ,
(4πDt)1/2
respectively.
126
this method in the context of electrostatics, where it yields the solution to Poisson’s
equation in the presence of specific boundary conditions, provided the problem has
a convenient symmetry. However, the method is quite general, and is applicable in
other instances as well. Broadly speaking, the method of images is a technique to find
the Green function of a linear differential operator with specific boundary conditions,
provided the problem has some symmetry that can be exploited. It is based on the
fact that the differential equation concerned, taken together with appropriate initial
and boundary conditions, has a unique solution. My purpose in using the method
of images in the present context of the diffusion equation is two-fold. The first is to
demonstrate how powerful the method is, when the conditions are right. The second
is to show that the method is also applicable to time-dependent problems, and not just
time-independent problems as in electrostatics.
Imagine placing mirrors at both boundaries, facing each other. A source point x
therefore has an infinite number of images, at the points −2b − x, 2b − x, −4b + x, 4b +
x, . . ., as we move away from the source point on both sides of the x-axis. (Draw a
figure and mark the successive image points.) The solution is a superposition of the
fundamental Gaussian solution with each of these points as the coordinate argument.
The boundary conditions above are incorporated by the following simple prescriptions:
(i) For RBC, the coefficient of the contribution from each image is just 1.
(ii) For ABC, the coefficient of the contribution from an image arising from n reflec-
tions is (−1)n .
The complete solutions in the two cases may therefore be written down. Show that,
after a bit of simplification, they are given by
1
∞
2
pref (x, t) = 1/2
e−(x+2nb) /(4Dt) (RBC at x = ±b)
(4πDt) n=−∞
and
1 ∞
n −(x+2nb)2 /(4Dt)
pabs (x, t) = (−1) e (ABC at x = ±b).
(4πDt)1/2 n=−∞
127
6. You may now ask: can each of these solutions be written as the fundamental Gaus-
sian solution on the infinite line, plus an ‘extra’ piece arising from the presence of the
finite boundaries? Show that the solutions above can be re-written in the form
) *
2
e−x /(4Dt) ∞
2 2
nbx
pref (x, t) = 1+2 e−n b /(Dt) cosh (RBC)
(4πDt)1/2 n=1
Dt
and
) *
2
e−x /(4Dt)
∞ nbx
−n2 b2 /(Dt)
pabs (x, t) = 1+2 (−1)n e cosh (ABC),
(4πDt)1/2 n=1
Dt
• It is remarkable that the only difference between the solutions for reflecting and
absorbing boundary conditions is the extra factor (−1)n in the summand in the
latter case. But this is sufficient to alter completely the long-time behavior of
the PDF p(x, t), as you will see below.
1 − n2 π2 Dt/b2
∞
1 nπx
pref (x, t) = + e cos (RBC)
2b b n=1 b
128
and
1 −(2n+1)2 π2 Dt/(4b2 )
∞
(2n + 1)πx
pabs (x, t) = e cos (ABC).
b n=0 2b
(a) Using the solution for pabs (x, t) with ABC, check that
verify that S(0, ±b | 0) = 1, as required. You will need the value40 of the
(b) Hence
series ∞ n
0 (−1) /(2n + 1).
39
More generally, the survival probability in any region should be computed by averaging over all
possible starting points as well, but I do not go into this detail here.
40
This was first deduced by the father of mathematical analysis, Madhava of Sangamagrama (1350-
1425), as a special case of the power series for tan−1 x discovered by him. The latter was rediscovered
a few centuries later by Gregory (1638-1675), and is now known as the Madhava-Gregory series.
41
Also called the escape time, or hitting time, or exit time, depending on the application.
129
Let Q(t, ±b | 0) dt be the probability for a particle starting from x = 0 at t = 0 to
reach either b or −b for the first time in the time interval (t, t + dt), without ever having
hit either of the end-points at any earlier time. It is obvious that the probability density
Q(t, ±b | 0) is nothing but the rate of decrease of the survival probability S(t, ±b | 0).
That is,
πD
∞
d 2 2 2
Q(t, ±b | 0) ≡ − S(t, ±b | 0) = 2 (−1)n (2n + 1) e−(2n+1) π Dt/(4b ) .
dt b n=0
∞
(a) Verify that Q(t, ±b | 0) is normalized to unity, i.e., 0
dt Q(t, ±b | 0) = 1.
Hence a first passage to either one or the other of the end-points is a sure event, as
already stated.
The mean first-passage time (MFPT) t(0 → ±b) is also easily determined. By
dimensional considerations, it must be proportional to b2 /D, which is the only time
scale in the problem.
(b) Show that the MFPT t(0 → ±b) = b2 /(2D), justifying the
term ‘diffusion time’
for this mean value. You will need the value of the sum ∞ n 3
0 (−1) /(2n + 1) ,
3
which is π /32.
Connection with the Schrödinger equation for a free particle: Let’s return
to the fundamental solution of the diffusion equation in d-dimensional space, for an
arbitrary initial PDF. This solution enables us to write down a similar solution for
another very important equation, namely, the Schrödinger equation for the position-
space wave function ψ(r, t) of a nonrelativistic free particle of mass m moving in d-
dimensional space:
∂ψ(r, t) i 2
= ∇ ψ(r, t).
∂t 2m
130
It is obvious that this equation has exactly the same form as the diffusion equation,
with the physically and mathematically important difference that the real diffusion
constant D in the latter is replaced by the pure imaginary constant i/(2m). We can
now write down the formal solution to the Schrödinger equation by analogy with the
solution of the diffusion equation. Given that the wave function is ψ(r, t ) at some
instant of time t , the solution at any time t > t is given by
d/2
m d im(r − r )2
ψ(r, t) = d r exp ψ(r , t ).
2πi(t − t )
2(t − t )
But we know from quantum mechanics that the solution must have the form
ψ(r, t) = dd r K(r, t ; r , t ) ψ(r , t ), (t > t )
where H is the free-particle Hamiltonian p2 /(2m), and |r is the position eigenstate
of the particle corresponding to the position eigenvalue r. Hence the explicit form
of the free-particle propagator for a nonrelativistic particle moving in d-dimensional
Euclidean space is
d/2
m im(r − r )2
K(r, t ; r , t ) = exp (t > t ).
2πi(t − t ) 2(t − t )
The replacement of the real constant D by the pure imaginary constant i/(2m) actu-
ally represents a drastic change. The kernel exp [−(r−r )2 /4D(t−t )] decays to zero as
|r−r | → ∞. On the other hand, the kernel exp [im(r−r )2 /2(t−t )] is an oscillatory
function with a modulus equal to unity. As a result, questions of convergence arise, and
these require careful handling. I do not go into these aspects here, but merely mention
that the formal similarity between the diffusion equation and the Schrödinger equation
has important consequences. The solution for ψ(r, t) written down above serves as the
starting point for the path integral formulation of quantum mechanics.
10. Spreading of a quantum mechanical wave packet: As you know from ele-
mentary quantum mechanics, the wave packet representing a free particle undergoes
dispersion, i.e., it broadens or spreads in time, even though the particle moves in a
vacuum. The physical reason for this fact is easy to see. Let ε and p denote the energy
and the magnitude of the momentum of the particle, respectively. Wave-particle dual-
ity is expressed by the Einstein-de Broglie relations ε = ω and p = k, where ω is the
angular frequency and k is the wave number. Hence the relation ε = p2 /(2m) becomes
ω = k 2 /(2m). This nonlinear dispersion relation immediately implies that the phase
velocity ω/k is not equal to the group velocity dω/dk, i.e., there is dispersion. In other
131
words, an initial wave packet will change shape and spread with time, even though the
particle is free, and not under the influence of any force.
In order to see this quantitatively, let’s consider the problem in one spatial dimen-
sion. Setting t = 0 for simplicity, we have
m 1/2 ∞ 2
ψ(x, t) = dx eim(x−x ) /(2t) ψ(x , 0).
2πit −∞
Therefore the initial uncertainty (i.e., the standard deviation) in the position of the
particle is ∆x(0) = σ.
(b) Show also that
∞
d
p(0) = dx ψ ∗ (x, 0) (−i) ψ(x, 0) = p0 ,
−∞ dx
while ∞
d2 2
p (0) =
2
dx ψ ∗ (x, 0) (−i)2 ψ(x, 0) = p 2
0 + .
−∞ dx2 4σ 2
Hence the initial uncertainty (or standard deviation) in the momentum is ∆p(0) =
/(2σ). It follows that ∆x(0) ∆p(0) = 12 . In other words, the initial state is a mini-
mum uncertainty state.
The Hamiltonian of a free particle is given by just the kinetic energy term, p2 /(2m).
As this is quadratic in the dynamical variable concerned, Ehrenfest’s Theorem im-
plies that the expectation value p(t) will remain equal to p0 for all t, while x(t) will
be given by p0 t/m: these are the expressions that would obtain for a classical particle.
These statements can be checked out directly, as follows.
132
(c) The integral representing ψ(x, t) now becomes a shifted Gaussian integral, and
can be evaluated, to find the wave function at any time t. Show that
eφ(x,t)
ψ(x, t) = ,
(2πσ 2 )1/4 [1 + (it/2mσ 2 )]1/2
where the exponent φ(x, t) is given by
imx2 im(x − p0 t/m)2 (x − p0 t/m)2
φ(x.t) = − − .
2t 2t[1 + (2 t2 /4m2 σ 4 )] 4σ 2 [1 + (2 t2 /4m2 σ 4 )]
The first two terms on the right-hand side of the last equation are purely imaginary,
and hence only contribute phase factors to ψ(x, t). They do not, therefore, affect the
probability density function |ψ(x, t)|2 . This quantity is given by
1/2
1 (x − p0 t/m)2
|ψ(x, t)| =
2
exp − .
2π[σ 2 + (2 t2 /4m2 σ 2 )] 2[σ 2 + (2 t2 /4m2 σ 2 )]
Thus, the PDF of the position of the particle at any time t > 0 is also a normalized
Gaussian, with its peak at x = p0 t/m, and a variance that increases with time.
(d) Show that
p0 t
x(t) = ,
m
2 2 t2
x2 (t) − x(t)2 ≡ ∆x(t) = σ 2 + .
4m2 σ 2
(e) Show also that
∞
p(t) = ∗ d
dx ψ (x, t) (−i) ψ(x, t) = p0 ,
−∞ dx
∞
d2 2
p (t) =
2
dx ψ ∗ (x, t) (−i)2 ψ(x, t) = p 2
0 + .
−∞ dx2 4σ 2
133
(iii) The uncertainty in the momentum at any time t remains equal to its initial value,
/(2σ). Thus the state of the particle is no longer a minimum uncertainty state
for any t > 0.
Observe, incidentally, that the energy of the particle (i.e., the expectation value of its
Hamiltonian) in the state under consideration is not p20 /(2m), but rather
1 p2 2
E ≡ H = p2 = 0 + .
2m 2m 8mσ 2
H, of course, is constant in time.
134
12 The Green function for (∇2 + k 2); nonrelativistic
scattering
The Helmholtz operator: After the Laplacian operator ∇2 and the diffusion op-
erator (∂/∂t − D∇2 ), we turn now to the Helmholtz operator ∇2 + k 2 , where k is a
constant. As you know, the normal modes of vibration of a region R are given by the
solutions of the Helmholtz equation (∇2 + k 2 )u(r, t) = 0, i.e., the eigenvalue equation
∇2 u = −k 2 u, with appropriate boundary conditions on the ‘displacement’ u at the
boundary S of the region. The most common boundary condition is the Dirichlet
boundary condition u(r, t) = 0 for r ∈ S. A vast literature exists on this problem
and its solutions.
Here, we shall not be concerned with the foregoing eigenvalue problem, but rather,
with the inhomogeneous equation (∇2 + k 2 ) f (r, t) = g(r, t). More specifically, we are
interested in finding the fundamental Green function for the operator ∇2 + k 2 . As
it is useful to work in the context of a specific physical application, I will consider
quantum mechanical scattering theory. For simplicity, we’ll focus on the elastic scat-
tering of a nonrelativistic particle of mass m from a static central potential V (r) that
vanishes as r → ∞. It is convenient (and conventional) to introduce a parameter λ as
a measure of the ‘strength’ of the potential, so that various physical quantities can be
expressed as power series in λ, and approximations to different orders in λ can be made.
E = 2 k 2 /(2m)
throughout. The scattered wave vector can be directed along any direction in space.
The quantity of interest is the probability of scattering in any particular direction. This
42
This technical difficulty can be overcome by working with normalizable wave packets rather than
plane waves as wave functions. But we shall not enter into this aspect here.
135
is measured by the differential cross-section dσ/dΩ, defined as
dσ flux of particles per unit solid angle in the direction concerned
= .
dΩ incident flux
Sufficiently far away from the scattering center (i.e., as r → ∞, or more precisely,
for kr 1), the scattered wave has the form of an outgoing spherical wave. The radial
coordinate dependence of this wave is (in three-dimensional space) eikr /r, and it is, in
general, modulated by an amplitude f relative to the unit amplitude of the incident
plane wave. The amplitude f is, in general, a function of the angular coordinates θ
and ϕ, as well as the energy E (or the wave number k). Without loss of generality, we
may choose spherical polar coordinates such that the polar axis is along the direction
of the incident wave vector k. Then, the spherical symmetry of a central potential
clearly implies that there can be no ϕ-dependence in the amplitude f . There remains
a nontrivial θ-dependence, because the presence of an incident wave vector k breaks
the spherical symmetry, leaving an axial symmetry about the direction of k. The
total wave function is a superposition of the incident and scattered wave functions. Its
asymptotic form given by
eikr
ψ(r) = ψinc (r) + ψsc (r) −−−→ eik·r + f (k, θ) ,
kr 1 r
The energy and angle-dependent factor f (k, θ) is called the scattering amplitude.
1. There is a direct relation between the differential cross-section and the scattering
∗ ∗
amplitude. incident current density is (ψinc ∇ψinc − ψinc ∇ψinc )/(2mi), which is easily
simplified to k/m. Its magnitude k/m is the incident flux. The scattered current
∗ ∗
density is given by (ψsc ∇ψsc − ψsc ∇ψsc )/(2mi). What we need here is the radially
outward scattered flux through a cone of solid angle dΩ located at any angular position
(θ, ϕ). This is given by
∗
∗ ∂ψsc ∂ψsc
ψsc − ψsc r 2 dΩ.
2mi ∂r ∂r
Using the asymptotic form above for ψsc , show that this quantity simplifies to
136
Hence the differential cross-section is given by
Suppose, for a moment, we treat the right-hand side λU(r)ψ(r) as a ‘source’ term
in an inhomogeneous differential equation. The general solution will be the sum of
the complementary function and the particular integral. The former is a solution of
the homogeneous equation (∇2 + k 2 ) ψ(r) = 0, chosen so as to satisfy the boundary
conditions. In the present case, the complementary function is just the incident wave
eik·r . The particular integral involves G(r, r ), the Green function for the Helmholtz
operator (∇2 + k 2 ). Thus
ψ(r) = e + λ d3 r G(r, r ) U(r ) ψ(r ).
ik·r
The presence of the unknown function ψ in the integrand in the last term makes this
an inhomogeneous integral equation for the wave function, rather than a solution for
this quantity.
2. Green function for the Helmholtz operator: The Green function satisfies the
equation
(∇2r + k 2 )G(r, r ) = δ (3) (r − r ).
Now, the operator ∇2r + k 2 , the the δ-function on the right-hand side, as well as the
free boundary condition (the vanishing of G as r → ∞) are all translation invariant
(i.e., invariant under the shift r → r − r ). Hence G will once again turn out to be a
function of R = r − r . Introducing the Fourier transform of G(R) and following the
same steps as in the case of ∇2 , we now get
1 eiq·R
G(R) = − d 3
q .
(2π)3 q2 − k2
137
I have used q for the Fourier transform variable conjugate to R, since k has already
been used for the incident wave vector. As usual, go over to spherical polar coordinates
in q-space, and choose the polar axis along the direction of the vector R. Carry out
the angular integrations to get
∞ ∞
1 q sin qR 1 q (eiqR − e−iqR )
G(R) = − 2 dq 2 = − dq .
2π R 0 q − k2 8π 2 iR −∞ q2 − k2
Note that the answer is a function of R alone, which is why G(R) has been written
as G(R). We must now use an appropriate i-prescription to avoid the singularities of
the integrand at the points q = ±k on the path of integration.
• This prescription will also help us select the correct Green function, i.e., the
one that corresponds to an outgoing spherical wave with an asymptotic behavior
∼ eikr /r.
The next step is to close the contour of integration in the complex q-plane by adding
a semicircle to the line integral. The contribution of this semicircle to the integral must
vanish as its radius tends to infinity. It must therefore lie (i) in the upper half-plane for
the term involving eiqR , and (ii) in the lower half-plane for the term involving e−iqR .
Moreover, we want G(R) to have an asymptotic behavior ∼ eikr /r as r → ∞. Hence:
(i) Only the pole at q = k must contribute to the term proportional to eiqR . There-
fore, the pole at q = k must be displaced into the upper half-plane.
(ii) Only the pole at q = −k must contribute to the term proportional to e−iqR .
Therefore, the pole at q = −k must be displaced into the lower half-plane.
3. It is easy to work out the Green functions obtained by adopting the three other
ways in which the poles at q = ±k can be given infinitesimal imaginary displacements.
Let be a positive infinitesimal, as usual.
138
(a) Suppose we displace the pole at +k into the lower half-plane in q, and the pole
at −k into the upper half-plane. (Draw a figure.) Verify that the Green function
then vanishes identically, i.e.,
1 eiq·R
− lim d 3
q = 0.
(2π)3
→0 q 2 − (k − i)2
(b) Similarly, suppose both poles are displaced into either the upper or the lower half-
plane. (Draw the corresponding figures.) Show that the corresponding Green
function is
1 eiq·R cos kR
− lim d 3
q =− .
3
(2π)
→0 (q ± i) − k
2 2 4πR
Exact formula for the scattering amplitude: Using the correct Green function,
the integral equation for scattering becomes
ik|r−r |
λ 3 e
ψ(r) = e −
ik·r
dr U(r ) ψ(r ).
4π |r − r |
We need to extract the asymptotic behavior of the wave function from this integral
equation, and identify the scattering amplitude. As r → ∞ (more precisely, for kr
1), we may replace the factor 1/|r − r | in the integrand by just r −1 . The implicit
assumption here is that the potential V (r ) decays rapidly for very large values of r ,
so that the contribution to the integral from the region in which both r and r are large
is quite negligible. In the pure phase factor exp (ik|r − r |), however, we need to be
more careful. The general field point r at which the wave function is being calculated
is the point at which the particle detector is located, in order to measure the flux per
unit solid angle in that direction. Thus the wave vector of the scattered particle, k ,
is in the same direction as r: that is, k er = k . We then have
1/2 2r · r 1/2
ik |r − r | = ik r 2 − 2r · r + r 2 ikr 1 −
r2
r·r
ikr 1 − 2 = ikr − ik er · r = ikr − ik · r .
r
Inserting these approximations in the integral equation for ψ(r), we get
eikr λ
ψ(r) → e −
ik·r
d3 r e−ik ·r U(r ) ψ(r ).
r 4π
Comparing this result with the asymptotic form of the wave function, it follows at once
that the scattering amplitude is given by the formula
λ
f (k, θ) = − d3 r e−ik ·r U(r ) ψ(r ).
4π
139
While this formula is exact, it is not of much use unless we already know the full wave
function ψ(r) at all points in space.
Scattering geometry and the momentum transfer: For scattering in any given
direction, the vector
Q = (k − k)
is the final momentum of the particle minus its initial momentum. That is, Q is the
wave vector corresponding to the momentum transfer associated with the scattering
process. Recall that the energy E is related to k = |k| = |k | according to E =
2 k 2 /(2m). The relation between k, Q and θ is also easy to derive. (Draw a figure.) It
is easily seen that
Q2 = 2k 2 (1 − cos θ) = 4k 2 sin2 12 θ , so that Q = 2k sin 12 θ .
Born series and the Born approximation: Going back to the inhomogeneous
integral equation for the wave function, we can solve it iteratively for sufficiently small
values of the ‘strength’ |λ| of the potential. We find
ik|r−r |
3 e
ψ(r) = e − (λ/4π) d r
ik·r
U(r ) eik·r
|r − r |
ik|r−r | ik|r −r |
3 3 e e
2
+ (λ/4π) d r d r U(r ) U(r ) eik·r
|r − r | |r − r |
+ ···
The precise conditions on the potential under which this is a convergent series are
discussed in textbooks on quantum mechanics, and I shall not go into this aspect here.
Very roughly speaking, the series solution is valid when the effect of the potential is
weak as compared to the kinetic energy of the incident particle. Substitution of the so-
lution above in the formula for f (k, θ) gives an expression for the scattering amplitude
as a power series in λ, called the Born series.
It is trivial to write down the scattering amplitude to first order in λ. All we have
to do is to approximate ψ(r ) in the integrand by the incident wave exp (ik · r ) itself.
This is called the (first) Born approximation. Denoting the scattering amplitude in
this approximation by fB (k, θ), we have
fB (k, θ) = −(λ/4π) d3 r e−iQ·r U(r ), where Q = k − k.
Hence the scattering amplitude in the Born approximation is, up to a constant fac-
tor, the (three-dimensional) Fourier transform of the potential, with the momentum
140
transfer wave vector Q playing the role of the variable conjugate to r. Now, eik·r =
eip·r/ = r | p is the position space wave function corresponding to the initial momen-
−ik ·r
tum eigenstate |p of the particle. 3 Similarly, e = e−ip ·r = r | p ∗ = p | r.
Using the completeness relation d r |r r| = I, we see that
λ mλ
fB (k, θ) = − p |U|p = − 2
p |V |p.
4π 2π
In other words:
• The scattering amplitude in the Born approximation is essentially the matrix el-
ement of the potential energy operator between the initial and final free-particle
momentum eigenstates.
4. The angular integration in the expression for fB (k, θ) is easily carried out. Show
that
2mλ ∞
fB (k, θ) = − 2 dr r sin (Qr) V (r),
Q 0
where Q is the magnitude of the momentum transfer vector Q.
show that
2mλV0 (sin Qa − Qa cos Qa)
fB (k, θ) = .
3 Q3
Hence show that the forward scattering amplitude in the Born approximation is fB (k, 0) =
(2mλV0 a3 )/(32 ).
6. The Yukawa potential: According to quantum field theory, the forces between
elementary particles arise from the exchange of other particles which are the quanta
of gauge fields. In the nonrelativistic limit of scattering from a static potential, such
a force generically reduces to a form called the Yukawa potential, given by
e−r/ξ
V (r) = ,
r
where ξ is a positive constant with the physical dimensions of length. It represents the
‘range’ of the potential, and is essentially the Compton wavelength of the exchanged
particle. The latter is given by h/(µc), where µ is the (rest) mass of the exchanged
particle. The functional form of the Yukawa potential arises in other physical contexts
as well. For instance, the screened Coulomb potential in a dielectric medium has
141
the Yukawa form.
Show that, in the Born approximation, the scattering amplitude for the Yukawa
potential is given by
2mλξ 2
fB (k, θ) = − 2 2 2 ,
(Q ξ + 1)
where Q is the magnitude of the momentum transfer vector, as usual. Hence show
that the total cross-section in the Born approximation is
4π 2 mλ2 ξ 2 4mEξ 2 + 2
σB (E) = .
E2 8mEξ 2 + 2
For very large values of the incident energy E, the total cross-section falls off like 1/E.
This is a general feature of scattering from a wide class of potentials. (This class,
however, does not include the Coulomb potential!)
(i) It is also the exact expression for the differential cross-section for scattering in the
Coulomb potential λ/r, because all the higher-order corrections to the first Born ap-
proximation happen to vanish in this instance.
(ii) Moreover, the expression is actually independent of Planck’s constant, and is ex-
actly the same as the differential scattering cross-section of classical particles in a
Coulomb potential.
(iii) The differential cross-section for forward scattering (θ = 0) diverges. So does the
total cross-section, as the divergence at θ = 0 is not integrable.
All these ‘minor miracles’ and drawbacks are related to the long-range nature of the
Coulomb potential. It turns out that the assumption of an initial plane wave state
(ψinc = eik·r ) is incompatible with a potential that decays as slowly as 1/r. The
scattering problem has to be re-done in this case. The Schrödinger equation in the
43
If we set ξ = h/(µc), then ξ → ∞ corresponds to µ = 0. Recall that electromagnetic interactions
are mediated by the exchange of photons, which are massless quanta.
142
presence of a Coulomb potential can be solved exactly, using parabolic cylindrical
coordinates. The solution involves confluent hypergeometric functions. The initial
state is found to be a plane wave that is amplitude-modulated by a factor of 1/r, and
phase-modulated by a term involving ln r.
143
13 The wave equation
The last standard equation of mathematical physics to which we turn our attention in
this course is the wave equation. To be specific, we’ll consider the fundamental Green
function of the wave operator, that corresponds to the causal, retarded solution of
the wave equation. This is the solution that is of direct physical significance for the
propagation of signals.
Formal solution for the causal Green function: The wave equation for a scalar
function (or signal) f (r, t) is given by
1 ∂2
− ∇ f (r, t) = g(r, t),
2
c2 ∂t2
where g(r, t) is a specified function of space and time that represents the source of
the signal, and c is the speed of the signal. Although I have used c to denote this
speed, what follows is not restricted to the propagation of light in free space. We’ll
first consider, formally, wave propagation in a general d-dimensional Euclidean space
of infinite extent,44 and then work out the explicit solutions for the cases d = 1, 2 and
3 that are of direct physical interest. Subsequently, I’ll also comment on what happens
when d > 3.
The crucial difference between the wave equation and Poisson’s equation, is, of
course, the relative minus sign between the time derivatives and the Laplacian. This
makes all the difference in the world (literally!), and is a reflection of the fact that the
spacetime we consider has d space-like dimensions and 1 time-like dimension. From
the mathematical point of view, the wave equation is a hyperbolic partial differential
equation, while Poisson’s equation is an elliptic equation, and the diffusion equation
is a parabolic equation.
The particular integral of the inhomogeneous wave equation is given by the integral
representation ∞
f (r, t) = dd r dt G(r, t ; r , t ) g(r , t ),
−∞
d
where d r is the volume element in d dimensions. The Green function G satisfies the
equation
1 ∂2
− ∇ 2
G(r, t ; r , t ) = δ (d) (r − r ) δ(t − t ),
c2 ∂t2
where δ (d) denotes the d-dimensional δ-function. In physical terms, this Green function
represents the signal at time t at the point r, arising when a sharply-pulsed point
source of unit strength is switched on at the point r at the instant of time t . We are
interested in the Green function that satisfies natural boundary conditions, i.e., G → 0
44
So that ∇2 stands for the Laplacian in d-dimensional Euclidean space.
144
as |r − r | → ∞. Further, we seek the causal Green function that vanishes identically
for all t < t , so that there is no signal anywhere before the source is switched on. This
is the principle of causality. Hence G must satisfy the conditions
(In a region of finite extent, in the presence of boundary conditions at finite values of
r, the dependence of G on r and r cannot be reduced in general to a dependence on
the difference r − r alone.) The quantity K(r − r , t − t ) is sometimes referred to as
the propagator. As you will see, causality imposes an even stronger constraint. Since
the propagation of the signal occurs with a finite speed c, a signal emanating from the
point r at time t cannot reach the point r till time t + |r − r |/c. This feature will
also emerge automatically in the solution for G.
R = r−r and τ = t − t ,
respectively. Then
1 ∂2
− ∇ 2
G(R , τ ) = δ (d) (R) δ(τ ),
c2 ∂τ 2
where ∇2 now stands for the Laplacian operator with respect to R. We must impose
the conditions G = 0 and ∂G/∂τ = 0 for all τ < 0, and the boundary condition G → 0
as R → ∞. Now express G as a Fourier integral, according to
∞
dd k dω i(k·R−ωτ )
G(R , τ ) = d
e G(k, ω).
(2π) −∞ 2π
Observe that we use a Fourier transform in both space and time, rather than a Fourier
transform with respect to r and a Laplace transform with respect to t. This is helpful
in imposing the initial conditions on G, as you will see. The inverse relation is of course
∞
G(k, ω) = d R d
dτ e−i(k·R−ωτ ) G(R , τ ).
−∞
Further, ∞
(d) dd k dω i(k·R−ωτ )
δ (R) δ(τ ) = e .
(2π)d −∞ 2π
145
Using these expressions in the partial differential equation satisfied by G(R , τ ), we
get ∞
dd k dω i(k·R−ωτ ) 2 2 2
e (ω − c k ) G(k, ω) + c = 0.
2
(2π)d −∞ 2π
But the set of functions {ei(k·R−ωτ ) }, where ω and the Cartesian components of k
run over all real values, forms a complete orthonormal basis in the space of integrable
functions of τ and R. Therefore the expression in curly brackets must vanish, for every
value of ω and the components of k. Hence
ω) = − c2
G(k, , where k 2 = |k|2 = k12 + . . . + kd2 .
(ω 2 − c2 k 2 )
As expected, the introduction of the Fourier transform has converted the partial differ-
ential equation for G into a trivially-solved algebraic equation for its Fourier transform
ω). Inverting the Fourier transforms, we obtain the formal solution for G(R , τ ),
G(k,
namely,
dd k ik·R ∞ dω e−iωτ
G(R , τ ) = −c 2
e .
(2π)d −∞ 2π (ω − c k )
2 2 2
We now encounter a familiar difficulty: the formal solution above does not make sense
as it stands! The integral over ω diverges because the integrand has poles at ω = −ck
and ω = ck that lie on the path of integration.
Now, for τ < 0, this semicircle must lie in the upper half-plane in ω, because it
is only in this region that the factor e−iωτ in the integrand vanishes exponentially as
45
You will recall that we’ve already used this sort of contour completion more than once—for in-
stance, in the derivation of dispersion relations for the generalized susceptibility in linear response
theory, and of the Green function corresponding to outgoing spherical waves for the Helmholtz oper-
ator.
146
Ω → ∞. The addition of the semicircle to the contour would then simply add a van-
ishing contribution to the original line integral that we want to evaluate. Therefore,
provided no singularities of the integrand lie on the real axis or in the upper half-plane
in ω, the integral over the contour C+ is guaranteed to vanish identically for τ < 0.
But this is precisely what is required by causality: namely, that G(R , τ ) be equal to
0 for all τ < 0.
On the other hand, for τ > 0, we do expect to have a signal that does not vanish
identically. But now the semicircle closing the contour must lie in the lower half-plane,
because it is only then that the factor e−iωτ in the integrand vanishes exponentially
as Ω → ∞; and hence so does the contribution from the semicircle to the integral
over the contour C− . Therefore, provided all the singularities of the integrand are
in the lower half-plane, all our requirements are satisfied. This is ensured by displac-
ing each of the poles of the integrand at ω = −ck and ω = +ck by an infinitesimal
negative imaginary quantity −i where > 0, and then passing to the limit → 0
after the integral is evaluated. Equivalently, we may replace ω by ω + i in the de-
nominator of the integrand, and take the limit → 0 after carrying out the integration.
dω e−iωτ
= lim .
Ω→∞ C 2π (ω + i)2 − c2 k 2
±
As explained in the foregoing, we must use C+ for τ < 0 and C− for τ > 0.
1. Show that
dd k sin cτ k ik·R
G(R , τ ) = c θ(τ ) e .
(2π)d k
Note how the step function θ(τ ) required by causality emerges automatically in the
solution for G(R , τ ).
2. Each of the two poles of the integrand in the formal solution for G(R , τ ) the
we started with can be displaced so as to lie either in the upper or lower half-plane.
This leads to four possible ways of making the divergent integral finite. The partic-
ular i-prescription we have used above is tailored to ensure that the correct causal
solution is picked up from among the set of possible solutions. Find the other solutions.
147
We are now in a position to consider the solutions for different values of d. In order
to make this explicit, I’ll denote the Green function by G(d) (R , τ ) instead of G(R , τ ),
from now on.
It is obvious from this expression that G(1) (X, τ ) is a symmetric function of X, i.e.,
(Change the variable of integration from k to −k, and the result follows.) As you will
see shortly, G(1) (X, τ ) is actually a function of |X|. Show that
where ε(x) = 1 for x > 0, and −1 for x < 0. Simplify this result to get
The factor θ(cτ − |X|) ensures that the signal arising from the source point x at time
t does not reach any point x until time t + |x − x |/c, as required by causality and the
finite speed of signal propagation. The presence of this step function makes the other
step function, θ(τ ), redundant from a physical point of view. However, it is present in
the formal mathematical solution for the quantity G(1) (X, τ ).
Another aspect of the solution is also noteworthy. An observer at x does not re-
ceive a pulsed signal, even though the sender sent out such a signal. (Recall that the
initial disturbance was a Dirac δ-function.) In fact, the signal received at any x persists
thereafter for all time. And it does so without diminishing in strength, a feature that
is unique to the case d = 1.
Let’s return to the general formula for G(R , τ ) in more than one spatial dimension,
i.e., d ≥ 2. The integrand on the right-hand side is a scalar, and so is the volume ele-
ment dd k. Further, the region of integration (namely, all of k-space) is invariant under
rotations of the coordinate axes. Hence G(d) (R , τ ) is a scalar, i.e., it is unchanged un-
der rotations of the spatial coordinate axes about the origin. This remains true for all
integer values of d ≥ 2. As a result of this rotational invariance, G(d) (R , τ ) is actually
a function of R and τ (where R = |R|, as usual). This means that we can choose the
148
orientation of the axes in k-space according to our convenience, without affecting the
result. This fact is of help in evaluating the integral. We’ll write the Green function
as G(d) (R, τ ), henceforth.
But we may regard this integral as the analytic continuation, to s = −icτ , of the
Laplace transform of J0 (kR), which is given by
∞
1
dk e−sk J0 (kR) = √ .
0 s2 + R 2
This analytic continuation can be justified properly, but I’ll not digress to do so here.
It is equivalent to putting in a decaying exponential factor (i.e., a regulator) like e−λk
in the integrand, and then passing to the limit λ → 0 after the integral is evaluated.
Show from the result above that
)
∞ 0 for c2 τ 2 < R2
Im dk eicτ k
J0 (kR) = √
0 1/ c2 τ 2 − R2 for c2 τ 2 > R2 .
A little care is needed in getting the overall sign right in the result. Note that the
Laplace transform, regarded as an analytic function of s, has branch points at s = ±iR,
and a branch cut running between these points. Identify the phases of the analytic
function (s2 + R2 )−1/2 at different points in the complex s-plane, and use this informa-
tion to arrive at the result quoted.
149
Since we are only interested in non-negative values of τ and R, the condition c2 τ 2 >
R yields a factor θ(cτ − R). We get, finally,
2
c θ(τ ) θ(cτ − R)
G(2) (R, τ ) = √ .
2π c2 τ 2 − R 2
The sharply-pulsed signal emanating from r at time t thus reaches any point r only
at time t + |r − r |/c, in accordance with causality and the finite velocity of propaga-
tion of the disturbance. But once again, the signal received at any field point r is no
longer a sharply pulsed one : it persists for all t > t + |r − r |/c, although its strength
slowly decays as t increases, like t−1 at very long times. Thus, both in d = 1 and in
d = 2, there is an after-effect even for a sharply-pulsed initial signal emanating from
the source.
The crucial point about the solution in three-dimensional space is this: If the source
pulse is a δ-function impulse emanating from r at time t , the signal at any field point
r is also a δ-function pulse that reaches (and passes) this point at precisely the instant
t + |r − r |/c. Hence there is no after-effect that lingers on at r, in stark contrast to
the situation in d = 1 and d = 2. Moreover, the amplitude of the pulse at r drops with
the distance from the source like 1/R, in exactly the way the Coulomb potential falls
off. These features are unique to three-dimensional space.
Another interesting reduction takes place in the case d = 3. Formally, if the limit
c → ∞ is taken in the wave equation, the wave operator reduces to the negative of the
150
Laplacian operator. We might therefore expect the solution for G(3) (R, τ ) to reduce to
the corresponding Green function for −∇2 . And indeed it does so, because the latter
Green function is precisely 1/(4πR) in three spatial dimensions.
where |r − r |
[g(r , t )]ret = g r , t −
def.
.
c
This is called the causal or retarded solution of the wave equation.
and
µ0 d3 r % &
A(r, t) = j(r , t ) ret .
4π |r − r |
Hence write down the formal solutions for the electric field E(r, t) and magnetic
field B(r, t) arising from arbitrary sources ρ(r, t) and j(r, t), using the relations E =
−∂A/∂t − ∇φ and B = ∇ × A. You must bear in mind that the solutions given
above for φ and A are, of course, only valid in the Lorenz gauge. But the expressions
obtained for the physical fields E and B are gauge-invariant, and hold good in any
gauge.
151
Interestingly enough, it turns out that the propagation of sharp signals is possible
in all odd-dimensional spaces with d ≥ 3, while it fails for all even values of d. This is
yet another manifestation of the fundamental differences that exist between Euclidean
spaces of even and odd dimensionalities, respectively. If a sharp δ-function pulse is
emitted at the origin at t = 0, then:
(i) The signal received at any point r is sharply pulsed, arriving and passing on at
time t0 + |r − r |/c with no after-effect, in spaces of dimension d = 3, 5, . . . .
(ii) In contrast, the signal lingers on for all t > t0 + |r − r |/c in spaces of dimension
d = 2, 4, . . . . We have already seen an example of such an after-effect in the case d = 2.
(iii) There is, however, one feature that is quite unique to d = 3: this is the only case
in which the original δ-function pulse is transmitted without any distortion, namely,
as a pure δ-function pulse.
There is an elegant and powerful way to solve the general problem. It is based on
the relativistic invariance of both the wave operator and the solution sought, i.e.,
on the fact that these remain invariant under rotations of the coordinate axes as well
as boosts to other inertial frames. (Here we have identified c with the speed of light in
a vacuum). I shall merely write down the final answer here. The causal Green function
turns out to be proportional to a derivative of a δ-function, according to
d(d−3)/2
G (R, τ ) ∝ (d−3)/2 δ(ξ 2 ),
(d)
where ξ = (c2 τ 2 − R2 )1/2 .
dξ
The order of the derivative, 12 (d−3), is an integer when d is an odd number. The Green
function remains a sharply-pulsed quantity in this case, although it is only for d = 3
that you get just a δ-function. For larger odd values of d (= 5, 7, . . . ), the fundamental
solution is given by higher and higher derivatives of the δ-function. These are increas-
ingly singular quantities. When d is an even integer, on the other hand, the solution is
a fractional derivative of a δ-function. Fractional derivatives are nonlocal objects,
defined in terms of suitable transforms (such as the Fourier transform). This is how
the extended nature of the Green function arises in the case of wave propagation in
even-dimensional spaces, leading to the after-effects mentioned earlier.
152
Dispersion and nonlinearity: Finally, I mention that there are two important ad-
ditional aspects of wave or signal propagation that can be adjusted so as to modify the
fundamental solution considered here. The first is dispersion, which occurs because
waves of different wavelengths propagate with different speeds in a medium. The cor-
responding dispersion relation (or frequency-wave number relationship) can be quite
complicated. The second aspect is nonlinearity. While the simple wave equation we
have considered here is linear in the signal f (r, t), physical situations often call for non-
linear equations. The interplay between dispersion and nonlinearity can be extremely
intricate and interesting, and a vast variety of new phenomena can arise as a result.
Among these are the so-called solitary waves and propagating solitons which rep-
resent very robust pulsed disturbances. These entities comprise a whole subject in
their own right.
153
14 The rotation group and all that
Rotations of the coordinate axes: Consider n-dimensional Euclidean space, where
n ≥ 2. A rotation of the coordinate axes about the origin of coordinates takes a
general point r to another point r , such that (i) the origin remains unchanged, and
(ii) the distance between any two points remains unchanged. Therefore each rotation
of the coordinate axes about the origin is a linear, homogeneous transformation of the
Cartesian coordinates. Such a transformation is specified by a (n × n) orthogonal
matrix R, as you’ll see shortly. That is, R satisfies the condition R RT = I, where
the superscript T denotes the transpose, and I is the unit matrix. Here’s one of the
simplest examples of a rotation matrix in 3-dimensional Euclidean space. The matrix
corresponding to a rotation of the coordinate axes about the origin, in the xy-plane,
and through an angle ψ, is given by
cos ψ sin ψ 0
R(ψ) = − sin ψ cos ψ 0 .
0 0 1
It is easy to check that R(ψ) is an orthogonal matrix. We could also have said, “a ro-
tation about the z-axis through an angle ψ”, because it is a rotation in 3-dimensional
space. In general, however, the correct way to specify a rotation in any number of
dimensions is to specify the plane in which the rotation takes place, rather than the
axis about which it occurs. This is because no such axis may exist in general, although
in n = 3 it so happens that a unique axis of rotation always exists for every rotation.
Basically, this is because the number of independent, mutually orthogonal axes (= n)
becomes equal to the number of independent, mutually orthogonal planes (= 12 n(n−1))
only for n = 3.
because the orthogonality condition on R implies that RT = R−1 . But we may also
regard components Tij of a tensor of rank 2 in n dimensions as the elements of a
(n × n) matrix T . Thus the transformation rule can be written in the compact form
T = R T R−1 . In other words, T is obtained from T by a similarity transformation
154
involving R.
The two expressions for r 2 must be equal to each other for every point in space.
Therefore we must have
For finite-dimensional square matrices, the left and right inverses are the same. Hence
RT R = I ⇒ R RT = I.
Proper and improper rotations: As we have just seen, a rotation of the coordi-
nate axes about the origin is specified (in n-dimensional Euclidean space) by an (n×n)
orthogonal matrix. These matrices form a group, denoted by O(n). Let R be a rota-
tion matrix, i.e., a matrix whose elements tell you what linear combinations of the old
coordinates yield the new coordinates. The orthogonality condition R RT = I on the
matrix R implies that (det R)2 = 1. Therefore det R = ±1.
Rotations for which det R = +1 are called continuous or proper rotations. They
are obtainable “continuously from the identity transformation”—that is, they can be
built up by a succession of infinitesimal rotations, starting from the identity transfor-
mation (or no rotation at all). For this reason, the set of proper rotations is called the
connected component of the rotation group. Proper rotations constitute a group
of their own, denoted by SO(n). (The ‘S’ stands for special, which means ‘unimodu-
lar’ or ‘with unit determinant’, in this context.) This is a subgroup of O(n). Proper
rotations preserve the orientation or handedness of the coordinate system. That is,
after a proper rotation, a right-handed coordinate system remains right-handed, and
a left-handed coordinate system remains left-handed.
155
In contrast, transformations with det R = −1 are called discontinuous or improper
rotations. They cannot be built up continuously from the identity transformation:
in general, they involve proper rotations together with reflections, such that a right-
handed coordinate system transforms to a left-handed one or vice versa. Examples of
such orientation-reversing transformations in n dimensions are:
(i) Reflection about any plane in space. (The plane need not be one of the planes
normal to the Cartesian axes.) In three dimensions, for example, a reflection
about the yz-plane corresponds to a transformation under which x → −x, y →
y, z → z.
156
are interested in the rotation matrices per se and in their algebraic properties. We
are also interested in finding the explicit transformation formula of a vector under an
arbitrary rotation of the coordinate axes. As you’ll see below, there’s quite an easy
way to arrive at the exact answer without any tedious algebra.
By cyclic permutation of xyz, we may write down the other two matrices
1 0 0 cos ψ 0 − sin ψ
R(ex , ψ) = 0 cos ψ sin ψ and R(ey , ψ) = 0 1 0 .
0 − sin ψ cos ψ sin ψ 0 cos ψ
It is easily checked that each of these matrices is orthogonal, and has a determinant
equal to +1. Hence each of them can be built up from the identity matrix by a succes-
sion of infinitesimal rotations about the axis concerned. We can work backwards from
the matrices written down above to see how this is done.
The parameter δψ has been factored out in the expression above. This makes the
elements of the matrix J3 pure numbers that are independent of the angle of rotation.
The reason for separating out the factor i in the definition of J3 is to ensure that J3 is
157
a hermitian matrix.48 The finite-angle rotation matrix R(ez , ψ) is then given by
where
0 0 0 0 0 i
J1 = 0 0 −i and J2 = 0 0 0 .
0 i 0 −i 0 0
The form R(ez , δψ) = I + i (δψ) J3 makes it quite clear why the matrix J3 is called
the generator of an infinitesimal rotation about the z-axis (i.e., a rotation about the
origin, in the xy-plane). Similarly, J1 and J2 may be identified as the generators of
infinitesimal rotations about the x and y axes (i.e., about the origin, in the yz-plane
and zx-plane), respectively.
158
(a) Work through the steps outlined above.
(b) Using the expressions given above for the matrices Jk , directly calculate the expo-
nentials eiψ Jk for k = 1, 2 and 3. (That is, sum the corresponding exponential se-
ries.) Verify that you recover the finite-angle rotation matrices R(ex , ψ), R(ey , ψ)
and R(ez , ψ).
(c) Verify that the generators Jk satisfy the commutation relations [Jk , Jl ] = iklm Jm .
2. The general rotation matrix in 3 dimensions: We’ve seen that the matrices
corresponding to rotations about the three Cartesian axes can be written as expo-
nentials of the corresponding infinitesimal generators. What about a general rotation
R(n, ψ) about an axis n pointing in an arbitrary direction in space?
In spite of this problem, it turns out to be possible to compute the exponential of the
(3 × 3) matrix i (J · n) ψ exactly, and in closed form. Here’s how this is done.
159
for the matrix elements of the rotation matrix R(n, ψ) is both simple and elegant. It
reads
Rij (n, ψ) = δij cos ψ + ni nj (1 − cos ψ) + ijk nk sin ψ.
Even more explicitly, if the spherical polar angles of the unit vector n are given by θ
and ϕ, we have
Using these expressions, you can write down the complete rotation matrix for an ar-
bitrary rotation R(n, ψ). Work out the steps outlined above, and find the explicit
expression for the rotation matrix R(n, ψ).
3. Find the eigenvalues and eigenvectors of the general rotation matrix R(n, ψ).
The finite rotation formula for a vector: Once we have R(n, ψ) explicitly, it is
straightforward to apply it to an arbitrary position vector r, to obtain
Expressing this formula back in terms of the vectors r and n helps us understand it in
physical terms:
It follows at once that, by its very definition, any vector a transforms under a general
rotation R(n, ψ) according to
R(n,ψ)
a −−−−→ a = (cos ψ) a + (1 − cos ψ) (a · n) n + (sin ψ) (a × n).
This is sometimes called the finite rotation formula for a vector. a has a com-
ponent along the original vector a, a component along the axis of rotation n, and a
component in the direction normal to the plane formed by a and n.
Relation between the groups SO(3) and SU(2): Orthogonal matrices and uni-
tary matrices are foremost among the sets of special kinds of matrices that form groups
under matrix multiplication, and that are of great importance in physics. All (n × n)
orthogonal matrices with real elements form the orthogonal group O(n), while all
(n × n) unitary matrices with complex elements form the unitary group U(n). When
the determinants of the corresponding matrices are further restricted to the value +1,
we have the special groups SO(n) and SU(n). It turns out that there is a very im-
portant connection between the rotation group SO(3) and the group of unimodular,
unitary (2 × 2) matrices, denoted by SU(2). This relationship has profound physical
consequences.
160
Let’s begin by observing that the three Pauli matrices σ1 , σ2 , σ3 (abbreviated as
σ) satisfy the same Lie algebra as the generators of rotations: if we set
Note that the matix (a · σ) uniquely determines the components (a1 , a2 , a3 ), and vice
versa. The corresponding transformation rule under a rotation, when we choose to
represent vectors by (2 × 2) matrices as above, is given by
where θ and ϕ, respectively, denote the polar and azimuthal angles specifying
the unit vector n.
161
(b) Hence verify that U is unitary and unimodular, i.e., UU † = I = U † U, and
det U = +1.
In other words, U(n, ψ) is an element of the special unitary group SU(2). Now, it is
obvious that replacing the matrix U(n, ψ) by −U(n, ψ) does not alter these properties.
Moreover, given any vector a, a rotation transformation made using −U in the place
of U leads to exactly the same a .
• Hence there are two distinct elements U(n, ψ) and −U(n, ψ) of SU(2), differing
only by an overall sign, corresponding to every rotation matrix R(n, ψ) ∈ SO(3).
Both the (2 × 2) unit matrix I and its negative −I are elements of SU(2). By them-
selves, these two matrices form a group, namely, the cyclic group of order 2, denoted by
Z2 .50 This group (Z2 ) is said to be the center of SU(2): its elements (I and −I) are
the only elements of SU(2) with which all the elements of the group commute. I and
−I constitute the kernel of the homomorphism from SU(2) to SO(3) : that is, they
are the only elements of SU(2) that map into the unit element of SO(3), the (3 × 3)
unit matrix. Thus there is an isomorphism between the quotient group SU(2)/Z2
and SO(3), written symbolically as
SU(2)/Z2 SO(3).
We’ll resume our discussion of other group theoretical properties after completing
an important bit of unfinished business: namely, the proof that the three generators
of rotations in 3-dimensional space themselves transform like a vector.
Recall that the actual representation of each Jk depends on what it is required to act
on. You’ve already come across two such representations: the defining representation
in terms of (3 × 3) matrices, and a (2 × 2) matrix representation in which Jk = 12 σk .
But the generators Jk operators can, and do, have an infinite number of other repre-
sentations, including infinite-dimensional ones—for example, when they act on state
vectors in infinite-dimensional Hilbert spaces in quantum mechanics. A rotation R of
the coordinate axes induces a transformation in the Hilbert space. This is a unitary
50
Z2 is also the group of integers under addition modulo 2: all the even integers are represented by
the identity element of Z2 , while all the odd integers are represented by the only other element of the
group.
162
transformation given by U(R) = ei (J·n) ψ , where the generators Jk (k = 1, 2, 3) have
the representation appropriate to the Hilbert space. What we need to do is to establish
that the triplet (J1 , J2 , J3 ) transforms like a vector under rotations of the coordinate
axes, independent of any particular representation for the generators. In effect, this
means that the only input we can use is the algebra satisfied by the three operators,
namely, [Jk , Jl ] = i klm Jm .
Let’s consider the general case of a rotation about an arbitrary axis n through an
angle ψ. (All special cases can then be read off from it.) The task is to show that J
satisfies the finite rotation formula, which is
e−i (J·n) ψ Ji ei (J·n) ψ = (cos ψ) Ji + (1 − cos ψ) (Jj nj ) ni + (sin ψ) ijk Jj nk .
Observe that the transformation rule for the operator J, given by the left-hand side of
this equation, is not just J = R(n, ψ) J. This is because J is not an ordinary vector,
but an operator-valued vector. I reiterate that the components of J are not necessarily
(3 × 3) matrices, nor do we care what the actual representation is. In vector form, we
need to show that
e−i (J·n) ψ J ei (J·n) ψ = (cos ψ) J + (1 − cos ψ) (J · n) n + (sin ψ) (J × n).
The right-hand side of this equation has precisely the form of the finite rotation for-
mula for the coordinate r of a point in three-dimensional space, with J replacing r.
Therefore, once the equation above is proved, we may assert that J itself transforms
like a vector under rotations.
What you will need for the purpose is a remarkable and extremely useful operator
identity called Hadamard’s Lemma. This identity expresses the operator eλA B e−λA ,
where A and B are (in general, non-commuting) operators acting on some linear vector
space and λ is a scalar constant (real or complex), as an infinite sum involving multiple
nested commutators of A with [A , B]. The formula is
5. Establish the transformation rule for J given above, using Hadamard’s Lemma.
Here’s an outline of the steps involved. Start with the identifications
λ = −iψ, A = J · n = Jk nk , B = Ji .
For ease of notation, write the r-fold multiple commutator in Hadamard’s Lemma as
Cr . Then
[A , Cr ] = Cr+1 , where C1 = [A , B] and r = 1, 2, . . . .
163
Using the commutator algebra [Jk , Jl ] = i klm Jm , it is straightforward to show that
C1 = i ilk Jl nk and C2 = [A , C1 ] = Ji − ni nk Jk .
Further,
Thus, C2r+1 = C1 and C2r = C2 . The infinite series in Hadamard’s lemma can now be
summed easily. The transformation formula for J follows upon simplification.
There’s a small subtlety involved in the left-hand side of the transformation for-
mula. The factor ei (J·n) ψ , which we have identified with R(n, ψ), appears on the right
of Ji , while its adjoint e−i (J·n) ψ appears on its left. You might have expected the
reverse, based perhaps on the transformation rule for a second-rank tensor. But the
order given here is correct. It is related to the fact that J itself is an operator (e.g.,
in a Hilbert space), and this is the transformation rule for an operator, as opposed to
that for a state vector.
λ2 % & λ3 % &
eλA B e−λA = B + λ [A , B] + A , [A , B] + A , A , [A , B] + · · · .
2! 3!
Here’s how the formula is derived. Define the operator-valued function the parameter
λ according to
F (λ) = eλA B e−λA .
Differentiate both sides with respect to λ, to get the differential equation satisfied by
F (λ):
F (λ) = A F (λ) − F (λ) A.
Note the ordering of the operators. A and F do not commute with each other, in
general. Now, F (λ) is a regular function of λ that can be expanded in a Taylor series
in powers of λ, with coefficients that are operators, of course. Thus
λ2
F (λ) = F (0) + λ F (0) + F (0) + · · · .
2!
The existence of such a power series expansion for F (λ) implies the existence of an
analogous series for its derivative F (λ) as well. This assertion follows from the fact
that F (λ) is an analytic function of the complex variable λ—in this instance, for all
finite values of |λ|. Hence
λ2
F (λ) = F (0) + λ F (0) + F (0) + · · · .
2!
164
Insert these series for F (λ) and F (λ) in the differential equation, and equate the coef-
ficients of like powers of λ. Note that F (0) = B. Determine the successive derivatives
F (n) (0) recursively, to arrive at Hadamard’s formula. This formula is the starting point
for the derivation of a number of extremely useful and important operator identities.
7. The general form of the elements of U(2) and SU(2): Let’s return to our
discussion of group theoretical aspects. You have seen that there is a homomorphism
between the rotation group SO(3) and the special unitary group SU(2). This is a
convenient place to determine the general forms of the elements of the unitary group
U(2) and the special unitary group SU(2).
More generally, an (n × n) matrix with complex elements has 2n2 independent real
parameters. Requiring that the matrix be unitary reduces this number to n2 . Thus
U(n) is an n2 -parameter group. The determinant of any element of U(n) is a complex
number of unit modulus, i.e., a number like eiθ . If, further, we require that the matri-
ces be unimodular, the number of independent real parameters is reduced further to
n2 − 1. Such matrices comprise the special or unimodular unitary group SU(n), which
is a subgroup of U(n).
165
The parameter spaces of SU(2) and SO(3): The parameter space of SU(2) follows
immediately: it is just the ‘surface’ of the unit (hyper)sphere
α12 + α22 + β12 + β22 = 1.
This is the 3-sphere S3 , which is simply connected: that is, any closed path in it can
be continuously deformed and shrunk to a point without leaving the space.
166
quantum number j. Tensors return to their original values under a 2π rotation of the
coordinate axes about any direction, while spinors change sign under such a rotation.
A rotation of 4π is needed to return a spinor to its original value.
The universal covering group of a Lie group: It turns out that every multiply-
connected Lie group G (that is, a Lie group with a parameter space that is not simply-
connected) has a unique universal covering group G whose parameter space is
simply connected. There is a homomorphism or many-to-one mapping from G to G.
• The Lie algebras of a Lie group and its covering group are, however, identical.
This means that the two groups are essentially identical in local neighborhoods (in
parameter space), although they may be very different from each other globally. For
one thing, the topologies of the parameter spaces are quite different from each other
(their connectivities are different). In the case of the rotation group SO(3), the param-
eter space (or group manifold) is doubly connected. The universal covering group
of SO(3) is the special unitary group SU(2). The latter is simply connected: its pa-
rameter space is the 3-sphere S3 . And, as I have already pointed out, there is a 2-to-1
homomorphism from SU(2) to SO(3): there are two matrices in SU(2), differing by
a sign, that correspond to each matrix in SO(3).
The group SO(2) and its covering group: These ideas are even more easily illus-
trated in the case of the group SO(2) of rotations 2-dimensional space. The parameter
space of SO(2) is a single angle modulo 2π. In other words, it is the 1-sphere S1 (the
circumference of a circle, in plain language). This space has an infinite number of
distinct classes of closed paths, labelled by the number of times the (directed) path
winds completely around the circle. When S1 is ‘rolled out’ on a line, the latter covers
the circle an infinite number of times, the copies being labeled by the elements of the
additive group of integers, Z. The universal covering group of SO(2) is therefore the
real number line R (which is also a Lie group!). We have the isomorphism
R/Z SO(2),
so that SO(2) is also a quotient group. Note also that SO(2) is equivalent to U(1),
the multiplicative group of complex numbers of unit modulus (i.e., complex numbers
of the form eiα , where α is real).
The groups SO(n) and Spin (n): What about the rotation group SO(n) for n =
4, 5, . . . ? It turns out that every one of these groups is also doubly connected, just as
SO(3) is. The universal covering group of SO(n) for n ≥ 3 is called Spin (n). There is
again a 2-to-1 homomorphism from Spin (n) to SO(n), and SO(n) is a quotient group
according to
Spin (n)/Z2 SO(n), n ≥ 3.
167
We’ve seen that Spin (3) SU(2). For higher values of n the covering group does not
reduce to anything so readily identifiable, except for the case of SO(6). In that case
it turns out that Spin (6) SU(4). These are the only two cases when Spin (n) is a
special unitary group.
Parameter spaces of U(n) and SU(n): For completeness, let me touch upon the
parameter spaces of the unitary group U(n) and the special unitary group SU(n),
where n ≥ 2.51 The parameter space of SU(n) is simply connected. U(n) is connected,
but not simply connected. The determinant of each member of U(n) is a complex
number of unit modulus. U(1) is a subgroup of U(n): matrices of the form eiα I
are included among the elements of U(n), and these matrices form a subgroup that
is isomorphic to U(1). The presence of this U(1) subgroup, whose parameter space
is the infinitely connected 1-sphere or circle S1 , makes the parameter space of U(n)
also infinitely connected. In technical terms, U(n) is the semidirect product of its
normal subgroup52 U(1) and its subgroup SU(n), and is written as
Conversely, the special unitary group SU(n) is just the unitary group U(n) ‘modulo’
U(1), i.e., it is a quotient group according to
U(n)/U(1) SU(n).
The universal covering group of U(n) is R ⊗ SU(n), which you can understand heuris-
tically because R is the universal covering group of U(1), and SU(n) is already simply-
connected.
A bit about the first homotopy group of a space: The connectivity of a space
V is formally determined by finding the different equivalence classes of directed
closed paths (or loops) in V. (Two such paths are equivalent if either of them can
be continuously and smoothly deformed into the other without leaving the space and
without cutting open the closed path.) More formally, each closed path is a map
S1 → V. The equivalence classes of such maps form a group, called the fundamental
group or the first homotopy group of V, denoted by π1 (V). If every possible closed
path in V can be shrunk continuously to a point, then π1 (V) reduces to the trivial group
with one element (this is often written as π1 (V) = 0), and V is a simply-connected
space. The converse is also true. Thus
168
The fundamental group of the 1-sphere S1 itself is the set of equivalence classes
of maps of S1 to S1 . This is π1 (S1 ) Z, the group of integers under addition. This
follows from the fact that the class of a directed closed path on S1 is labeled by the
winding number of the path (the number of times it goes around S1 completely). It is
intuitively clear that π1 (S2 ) = 0.53 It can be shown rigorously that π1 (Sn ) = 0 for all
n ≥ 2. As you might expect, since the 2-torus T2 is equivalent to the direct product
space S1 ⊗ S1 , we have π1 (T2 ) = Z ⊗ Z. This generalizes in an obvious fashion to the
n-torus Tn = S ⊗ ·
1
· · ⊗ S1 .
n factors
The fundamental groups of various group manifolds (and of other spaces as well,
of course) are of great interest. Based on what has been mentioned in the foregoing,
here are some of the results concerned.54
π1 (Rn ) = 0, n ≥ 1.
π1 (S1 ) = π1 (T1 ) = Z.
π1 (Sn ) = 0, n ≥ 2.
π1 (Tn ) = π1 (S1 ⊗ · · · ⊗ S1 ) = Z ⊗ · · · ⊗ Z.
π1 (SO(2)) = π1 (U(1)) = π1 (S1 ) = Z.
π1 (Spin (n)) = 0, n ≥ 3.
π1 (SO(n)) = π1 (Spin (n)/Z2 ) = Z2 , n ≥ 3.
π1 (SU(2)) = π1 (S3 ) = 0.
π1 (SU(n)) = 0.
π1 (U(n)) = π1 (U(1) SU(n)) = Z.
53
A fact that is picturesquely expressed by the statement, “You can’t lasso a basketball!”
54
When we write π1 (G) where G is a Lie group, we mean the fundamental group of the parameter
space of G.
169
15 QUIZ 2
I. Are the statements in quotation marks true or false?
1. “The function sin (1/z) does not have a Taylor series expansion in the neighbor-
hood of z = 0.”
∞
2. A function f (z) is defined by the power series z 2n+1 /[n! (n + 1)!] about the
n=0
origin.
∞
4. “The Mittag-Leffler expansion of the gamma function is given by Γ(z) = (−1)n /[(z+
n=0
n) n!].”
∞
6. “The power series (ln n) z n /n converges at all points on its circle of conver-
n=1
gence.”
8. “Dispersion relations for the real and imaginary parts of a generalized suscepti-
bility χ(ω) can be derived only if the corresponding response function φ(t) decays
to zero faster than any negative power of t, as t → ∞.”
9. “The derivative of the gamma function, Γ (z), has zero residue at each of its
poles.”
170
10. “The Legendre function of the second kind, Qν (z), has branch points in the z-
plane even when ν is a positive integer.”
11. “The Laplace transform of the function f (t) = cosh πt has no singularities in the
region Re s > π.”
d2 d
12. Bessel’s differential equation is z 2 2 + z + (z 2 − ν 2 ) f (z) = 0, where ν is
dz dz
a parameter.
“If φν (z) is any solution of this equation, then φ−ν (z) must be equal to φν (z),
apart from a possible multiplicative constant.”
13. “The group Möb (2, C) of Möbius transformations of the complex plane has con-
tinuous subgroups, but no discrete subgroups.”
14. “The group Möb (2, C) of Möbius transformations of the complex plane is iso-
morphic to the group SO(3, 1) of homogeneous proper Lorentz transformstions
in (3 + 1)-dimensional spacetime.”
15. “The Riemann surface of the function f (z) = z 1/2 (z − 1)−1/3 has 6 sheets.”
16. “It is possible to find a contour integral representation of the beta function
B(z, w) that is valid for all complex values of both z and w.”
17. “The Riemann zeta function ζ(z) cannot be continued analytically to the left of
the line Re z = 12 , because it has an infinite number of zeroes on that line.”
18. “The Fourier transform operator in L2 (−∞ , ∞) has a finite number of eigenval-
ues, each of which is infinitely degenerate.”
19. Let G(x, x ) denote the Green function of the differential operator d2 /dx2 where
x ∈ [−1, 1].
171
20. “The fundamental Green function of the Laplacian operator ∇2 in four-dimensional
Euclidean space is G(r, r ) = −1/[4π 2 (r − r )2 ].”
21. Consider the diffusion equation in d-dimensional space, ∂f /∂t = D∇2 f with
boundary condition f (r, t) → 0 as r → ∞ and initial condition f (r, 0) = δ (d) (r).
22. The scattering amplitude for the scattering of a nonrelativistic particle of mass
m in a central potential λV (r) is given by
mλ
f (k, θ) = − 2
d3 r e−ik · r V (r) ψ(r),
2π
where k is the scattered wave vector.
“This formula is valid only if the potential V (r) decays to zero as r → ∞ more
rapidly than any inverse power of r.”
23. Continuation: “In the Born approximation, the scattering amplitude in the for-
ward direction (θ = 0) vanishes identically.”
24. Continuation: “In the Born approximation, the imaginary part of the scattering
amplitude vanishes identically.”
172
28. Let J = (Ji , J2 , J3 ) be the generators of rotations in three-dimensional space,
satisfying the Lie algebra [Jj , Jk ] = ijkl Jl .
30. “The first homotopy group of the parameter space of the special orthogonal group
SO(n), for every n ≥ 3, is Z2 .”
3. The region of absolute convergence in the complex z-plane of the power series
∞ %
&
(n + 1)/(n2 + 1) ( 12 z)n is · · · .
n=0
5. Let [z1 , z2 ; z3 , z4 ] denote the cross-ratio of the four points z1 , z2 , z3 and z4 in the
complex plane. Then [z1 , z2 ; z3 , z4 ] + [z1 , z3 ; z2 , z4 ] = · · · .
173
∞
8. The Bernoulli numbers Bn are defined via the expansion z/(ez −1) = Bn z n /n!.
n=0
Therefore Bn is given by the contour integral Bn = · · · . (You must specify both
the integrand and the contour.)
9. The Chebyshev polynomial of the second kind, Un (cos θ), has the generating
function
1 ∞
= Un (cos θ) tn ,
1 − 2t cos θ + t2
n=0
10. Continuation: Evaluating the contour integral and simplifying the result, the
final expression for Un (cos θ) is Un (cos θ) = · · · . (You must express your answer
in terms of trigonometric functions of θ.)
is f(s) = · · · .
16. Let
1 − |x|, |x| ≤ 1
f (x) =
0, |x| > 1.
If f(k) denotes the Fourier transform of f (x), the value of the integral
∞
dk |f(k)|2 = · · · .
−∞
174
where λ is the mean jump rate and pi , qi are directional probabilities such that
p1 + q1 + p2 + q2 = 1. The leading asymptotic behavior of P (, m, t) at very long
times (λt 1) is given by P (, m, t) ∼ · · · .
18. The diffusion equation for the positional probability density of a particle diffusing
on the x-axis in the region −∞ < x ≤ a, in the presence of a constant force, is
given by
∂p(x, t) ∂p (x, t) ∂ 2 p (x, t)
= −c +D .
∂t ∂x ∂x2
Here c and D are positive constants denoting the drift velocity and diffusion
constant, respectively. p(x, t) is normalized to unity for all t ≥ 0. There is a
reflecting boundary at the point x = a. The boundary condition satisfied by
p(x, t) at x = a is then given by · · · .
20. Continuation: The normalized solution for pst (x) is pst (x) = · · · .
21. A quantum mechanical particle of mass m moving in one dimension has the
Hamiltonian H = p2 /(2m), where p is the momentum operator of the particle.
Its momentum-space wave function at t = 0 is given to be φ(p). Therefore its
momentum-space wave function at any time t ≥ 0 is given by ψ(p, t) = · · · .
24. Let R = r − r and τ = t − t , as usual. Let G(d) (R, τ ) denote the fundamen-
tal Green function of the Klein-Gordon operator + µ2 , where µ is a positive
constant and = (1/c2 )(∂ 2 /∂t2 ) − ∇2 , in (d + 1)-dimensional spacetime. Then
G(d) (R, τ ) can be expressed in the form
(d) 1
G (R, τ ) = dd k φ(k, R, τ ),
(2π)d
where φ(k, R, τ ) = · · · .
175
25. Let J = (J1 , J2 , J3 ) denote the generators of rotations in three-dimensional
space, and let ψ be an arbitrary angle. The quantity e−iJ1 ψ J2 eiJ1 ψ , expressed
as a linear combination of the generators, is · · · .
27. The number of generators of the orthogonal group O(n) and the special orthog-
onal group SO(n) are, respectively, · · · and · · · .
28. The number of generators of the unitary group U(n) and the special unitary
group SU(n) are, respectively, · · · and · · · .
29. Let x and p denote the position and momentum operators of a quantum mechan-
ical particle moving in one dimension, so that their commutator [x , p] = iI,
where I is the unit operator. Let a be a real constant with the physical dimen-
sions of length. Using Hadamard’s Lemma, the operator eiap/ x e−iap/ simplifies
to
eiap/ x e−iap/ = · · · .
This result tells us why the momentum operator p is the generator of translations
in position space.
30. Continuation: Let b be a real constant with the physical dimensions of linear
momentum. Once again, using Hadamard’s Lemma, the operator e−ibx/ p eibx/
simplifies to
e−ibx/ p eibx/ = · · · .
This result tells us why the position operator x is the generator of translations
in momentum space.
176
Quiz 2: Solutions
I. True or false:
1. T
2. T
3. F
4. F
5. T
6. F
7. T
8. F
9. T
10. T
11. T
12. F
13. F
14. T
15. T
16. T
17. F
18. T
19. T
20. T
21. T
22. F
23. F
24. T
177
25. F
26. F
27. T
28. T
29. F
30. T
2. m = tan ϕ and c = 0.
3. |z| < 2.
5. 1.
(w − w2 )(w1 − w3 ) (z − z2 )(z1 − z3 )
6. = .
(w − w3 )(w1 − w2 ) (z − z3 )(z1 − z2 )
7. λ = 16.
n! dz
8. Bn = , where C encloses the origin once in the positive sense.
2πi C (ez − 1)
zn
1 dt
9. Un (cos θ) = , where C encloses the origin once in
2πi C tn+1 (1 − 2t cos θ + t2 )
the positive sense.
sin (n + 1)θ
10. Un (cos θ) = .
sin θ
11. U1 (cos θ) = 2 cos θ.
√ √
12. i 2, −i 2, ∞.
14. t et .
16. 4π/3.
178
% √ √ &
(p1 /q1 )/2 (p2 /q2 )m/2 exp − λt 1 − 2 p1 q1 − 2 p2 q2
17. P (, m, t) ∼ .
4πλt (p1 q1 p2 q2 )1/4
∂p
18. D −cp = 0.
∂x x=a
26. 0.
27. 1
2
n(n − 1) and 12 n(n − 1).
28. n2 and n2 − 1.
29. x + a.
30. p + b.
179