Phép Biến Đổi Tích Phân Số Phức
Phép Biến Đổi Tích Phân Số Phức
Semester 2, 2022/23
4 Fourier transforms 39
4.1 Derivation of the Fourier transform from the complex Fourier series . . . . . . . . . 39
4.2 Complex Fourier transforms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
4.3 Proofs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
4.4 Fourier transform pairs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
4.5 The Dirac delta function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42
4.6 The δ-function as a limit of sharply peaked functions . . . . . . . . . . . . . . . . . 43
4.7 Fourier transforms in two or more dimensions . . . . . . . . . . . . . . . . . . . . . 43
4.8 Application of Fourier transforms to partial differential equations . . . . . . . . . . 44
4.9 Fourier sine and cosine transforms . . . . . . . . . . . . . . . . . . . . . . . . . . . 46
4.10 Odd and even extensions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
4.11 Application of sine and cosine transforms to PDEs . . . . . . . . . . . . . . . . . . 49
5 Laplace transforms 51
5.1 Properties . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52
5.2 Proofs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
5.3 Uses . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
5.4 The complex inversion theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55
5.5 Using Laplace transforms to solve partial differential equations . . . . . . . . . . . 60
5.6 Applications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64
6 Hankel transforms 65
6.1 The Bessel equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
6.2 Bessel functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
6.3 The Hankel transform . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67
6.4 Relation between FT and HT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67
6.5 Application of the Hankel transform to PDEs . . . . . . . . . . . . . . . . . . . . . 67
2
7 Nyquist stability theory 69
7.1 Stability of circuits . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69
7.2 Closed loop feedback circuits . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69
7.3 Nyquist’s stability criterion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70
1 Introduction
An integral transform relates an input function f (t) to an output function T f (u) by the relation
Z t2
T f (u) = dtK(t, u)f (t) (1)
t1
where K(t, u) is a function of two variables, known as the kernel of the transform. There are
many useful choices of K(t, u) and each choice defines a different integral transform. The functions
(f (t), T f (u)) are sometimes referred to as integral-transform pairs. The introduction of such a
transform in a given problem may be advantageous if T f (u) can be determined more easily than
f (t); for example, f (t) may satisfy a differential equation while T f (u) may satisfy an algebraic
equation. Most useful kernels have an associated inverse kernel K −1 (u, t) which gives the inverse
transform Z u2
f (t) = duK −1 (u, t)T f (u). (2)
u1
A symmetric kernel is one for which the kernel and inverse kernel coincide.
The most commonly used integral transforms are infinite, i.e. the range of integration is infinite,
and the most popular integral transforms are Laplace transforms; Fourier transforms (also called
complex Fourier transforms or exponential Fourier transforms); Fourier sine and cosine transforms
and the Hankel transform. We will consider all these transforms - their properties and their
applications - in this module. For example, the Laplace transform is defined as
Z ∞
f¯(s) = f (t)e−st dt, (3)
0
i.e. the integral over t extends over the positive real axis. In this case the kernel is
Note that notations for integral transforms vary widely; there are no universal conventions for the
labels of the arguments of the kernels, and notations for the transform itself also vary. Engineers
typically use (s, t) for the Laplace transform, with t being interpreted as a time coordinate, and
(k, x) for the Fourier transform, with x interpreted as a spatial coordinate. However one can
equally denote the Laplace transform as
Z ∞
f¯(k) = f (x)e−kx dx, (5)
0
3
Within the integer numbers (x ∈ Z) one can solve 3 + x = 2, but not 2x = 3;
Within the rational numbers (x ∈ Q) one can solve 2x = 3, but not x2 = 2;
Within the real numbers (x ∈ R) one can solve x2 = 2, but not x2 = −2;
Within the complex numbers (x ∈ C) one can solve any polynomial equation.
The fundamental theorem of algebra: Every (nonconstant) polynomial with complex co-
efficients has at least one zero in C.
Corollary: A polynomial of order n has exactly n zeros (roots) in C. Roots may occur with
multiplicity, i.e. the roots may be coincident.
Argand diagrams The relationship between Cartesian and polar representation of complex
numbers becomes clear on an Argand diagram - see Figure 1. This represents the complex plane
by a Real and an Imaginary axes (horizontal and vertical respectively, referring to x and y).
Im
z
r sin θ = y
=x+iy
=r ei θ
Re
θ
r cos θ = x
=x−iy
=r e−i θ
z
4
From the definitions it follows that
Im
open disk
a1
r1
closed disk
a2
circle r2
|z−a|=r
a
r
a3
r3
punctured open disk
Re
Disks in the complex plane A circle of radius r with centre a in the complex plane has the
Eq.
|z − a| = r, (12)
where a can be complex and r is real and positive. The inside of the circle
|z − a|< r, (13)
|z − a|≤ r. (14)
If we remove the point a from the disk it is called a punctured disk - such a disk is usually open:
f (z + h) − f (z)
f 0 (z) := lim (17)
h→0 h
5
Im
Re
Figure 3: In this definition the limit must be independent of the direction in which the complex
number h approaches zero.
exists. When it exists, the limit is called the derivative of the function f at z, and is denoted by
f 0 (z). To calculate the complex derivative of f (z) we need to consider
h = δx + iδy. (18)
Notice that for the limit in Eq. (17) to exit, it must be independent of the direction in which h is
taken to zero.
If the derivative of f exists everywhere in a domain D, then f is called holomorphic on D.
A function which is holomorphic on C is called entire.
Cauchy-Riemann equations Let f (z) be holomorphic in a certain domain, and let z = x + iy.
It is often convenient to write
f (z) = u(x, y) + iv(x, y), (19)
thereby splitting the complex function f into real and imaginary parts just like z. Then using (18)
f (z + h) − f (z) u(x + δx, y + δy) + iv(x + δx, y + δy) − u(x, y) − iv(x, y)
= (20)
h δx + iδy.
Take first δy = 0 (so, approach with h to zero along the x axis). Then h = δx, so that
f (z+h)−f (z) u(x + δx, y) + iv(x + δx, y) − u(x, y) − iv(x, y)
limh→0 h = limδx→0
δx
u(x + δx, y) − u(x, y) v(x + δx) − v(x, y) (21)
= limδx→0 + i
δx δx
6
Taking the real and imaginary parts of f 0 (z) and equating them we obtain the Cauchy-Riemann
equations:
ux = vy , uy = −vx . (24)
Now assume that u(x, y) and v(x, y) are twice differentiable, and that their first and second
derivative are continuous. (In fact, as we shall see later, if f (z) is holomorphic, u and v are
automatically guaranteed to have continuous partial derivatives of all orders!) By taking another
derivative of the Cauchy-Riemann equations, we obtain
and hence
uxx + uyy = 0, vxx + vyy = 0. (26)
We find that u(x, y) and v(x, y) both satisfy Laplace’s equation. Such functions are called
harmonic.
Conversely, it can be shown that if u and v obey the Cauchy-Riemann equations, then f = u+iv
is holomorphic. Hence:
u = x2 − y 2 v = 2xy
ux = 2x vy = 2x (27)
uy = −2y vx = 2y
7
2. At any point in D, f (z) can be written as
∞
X f (n) (a)
f (z) = an (z − a)n , where an = (29)
n=0
n!
and f (n) denotes the n-th derivative of f . This is called the Taylor series of f (z) about
z = a. The Taylor series converges (uniformly) anywhere in D.
Radius of convergence The region of convergence of the Taylor series is always a disk, so that
there is some R for which the series converges for all |z −a| < R, and diverges for all |z −a| > R. (If
f is analytic in the whole complex plane then R becomes ∞.) R is called the Radius of Convergence
(RoC). It can be thought of as the largest possible r for which (29) is true, or, equivalently, the
distance to the nearest point (from a) where f is no longer analytic.
1
Example: Find the Taylor Series of f (z) = 1−z about z = 0 and state its RoC.
Since f is analytic at z = 0 (can you prove that?), we can choose to calculate its derivatives at this
point by taking the limit along the real axis (as the direction does not matter). This is convenient,
since we can then treat f as a function of a real variable x, and calculate the derivatives of f (x)
in the usual manner: f 0 (z) = (1 − z)−2 , f 00 (z) = 2(1 − z)−3 , and so on. We find that the standard
binomial expansion applies:
∞
X
f (z) = (1 − z)−1 = zn. (31)
n=0
We know that this converges along the real axis for |x| < 1. Since the region of convergence must
be a disk, we conclude that the expansion converges on the complex plane for |z| < 1. Hence,
RoC=1.
1
Example: Find the Taylor Series of f (z) = 1−z about z = 3 and state its RoC.
Again, f is analytic at z = 3, and we can use the binomial expansion (this time with respect to
the variable z − 3) to obtain
−1 ∞ n
−1 −1 1 3−z 1 X 3−z
f (z) = (1 − z) = (−2 − (z − 3)) =− 1− =− (32)
2 2 2 n=0 2
Binomial Expansion The Binomial series often offers the quickest way in deriving power
series:
k(k−1) 2 k(k−1)(k−2) 3 k(k−1)(k−2)...(k−(n−1)) n
(1 + z)k = 1 + kz + 2 z + 3! z + ... + n! z + ...
∞ (33)
X k(k − 1) · · · [k − (n − 1)] n
= 1+ z which converges for |z| < 1.
n=1
n!
Many functions have well known power series, and hence RoC’s. These are the same as for their
real counterparts, with x replaced by z. In all cases the RoC about any point a is determined by
the location of “problem points”: these are the singularities of f (see Section 2.6).
8
3
Im f=1 / (1−z)
2 RoC=2
1 RoC=1
−2
−3
−2 −1 0 1 2 3 4 5 6
1
Figure 4: We might expect the function f (z) = 1−z not to be analytic at z = 1.
Im
Re
Figure 5: A Taylor series exists inside an analytic disk. A Laurent series exists inside an analytic
annulus. Singularities are indicated by crosses.
If z = a is itself a singularity then there is no Taylor series, since the disk D = {z : |z − a| < 0}
is empty. If z = a is a singularity and R is the distance from a to the nearest singularity, then a
Laurent series exists on the punctured open disk 0 < |z − a| < R.
9
1
Example: Find the Laurent series of f (z) = 1−z about z = 0 and state its annulus of convergence.
We found previously that f had a Taylor series about z = 0 within D = {z : |z| < 1}, i.e., with an
RoC of 1. This was due to a singularity at z = 1. We might expect a Laurent series of f about
z = 0 to be valid for {z : 1 < |z|}, if there are no other singularities of f (z). Using the binomial
expansion again we find
1 −1
1 1
f (z) = 1−z = z1 −1 = − z
z
1 − z1
n n+1 (35)
1 m
P∞ P∞ P∞ P−1
− z1 n=0 z1 = − n=0 z1
= = − m=1 z =− n=−∞ zn.
The binomial expansion is valid for z1 < 1, i.e., the Laurent series is valid in the (semi-infinite)
annulus A = {z : 1 < |z|}.
Note that this Laurent series has negative powers of z, unlike a Taylor series; in general, of course,
the Laurent series has both positive and negative powers.
2.6 Singularities
A point a is a regular point of f (z) if f is analytic at a. The point a is a singularity of f (z) if
a is a limit point of regular points which is not itself regular.
The point a is an isolated singularity if f (z) is analytic on the punctured disk {z : 0 <
|z − a| < r} for some r > 0.
If there exists no r > 0, however small, such that f (z) is analytic on {z : 0 < |z − a| < r}, then
a is called a non-isolated essential singularity, but we will not be dealing with these in depth
here.
Removable singularity The point a is a removable singularity if the principal part of the
Laurent series about a is zero.
Example: The Laurent series for z1 sin z about z = 0.
z −1 sin z = z −1 z − z 3 /6 + . . . = 1 − z 2 /6 + . . .
(36)
P∞ n
P∞ n
So n=−∞ cn z = n=0 cn z because cn = 0 for all n < 0; the principal part of the Laurent
P−1
series is zero, i.e., = n=−∞ cn z n = 0.
As power series are unique this is equivalent to the Taylor series.
10
Essential singularity The point a is an isolated essential singularity if the principal part
of the Laurent series has an infinite number of terms.
Example: The Laurent series for e1/z about z = 0.
X ∞ n X ∞ 0
1 1 1 1 −n X 1 m
exp = = z = z (37)
z n=0
n! z n=0
n! m=−∞
n!
The values of n for which c−n 6= 0 form an infinite sequence.
Poles The point a is a pole of order k if c−k 6= 0 and cn = 0 for all n < −k.
Since k is lowest power of (z − a) we have
∞
X
f (z) = cn (z − a)n . (38)
n=−k
A pole of order 1 is called a simple pole. A pole of order 2 is called a double pole. Poles of order
three and higher are usually referred to as such, although can be labeled triple etc.
2.7 Residues
The coefficient c−1 is called the residue of f at a and is denoted by Res(f : a).
Note that a removable singularity has no residue, since c−1 = 0 as part of c−1 = c−2 = c−3 =
. . . = 0. Both poles and essential singularities have residues since c−1 6= 0 in general, although we
can have a residue of zero in these cases (but recall that non-removable singularities would always
contain at least one non-zero negative coefficient in their Laurent expansion).
It is important for us to calculate residues quickly as they are used a great deal in complex
integration.
If a is a pole of order k, then its residue is
(k−1)
1 d
Res(f : a) = lim (z − a)k f (z) (pole of order k only) (42)
(k − 1)! z→a dz
(Proof by calculation from the Laurent series.) For a simple pole (k = 1) this general formula
reduces to
Res(f : a) = lim (z − a)f (z) (simple pole only). (43)
z→a
11
A special case for simple poles: For a simple pole of the form
g(z)
f (z) = , (46)
(z − a)
where g is nonzero and analytic at a we have
12
2.8 Contour integration, Cauchy theorem and residue theorem
Contours We are used to integrating a real function along the real line: considering the value of
f (x) as x ranges between α and β say and finding the area underneath.
For integration over a complex variable z we move in two dimensions across the complex plane.
Because of this it is not only the end points that are important but also the route across the
complex plane between them.
A curve γ is the graph of a continuous function z(t) from a real interval a ≤ t ≤ b to the
complex plane:
γ = {z(t) : a ≤ t ≤ b}. (56)
We will call z(t) together with the interval [a, b] the parameterised curve.
If z(a) = z(b) then γ is closed.
If z(t1 ) = z(t2 ) only if t1 = t2 for all t ∈ (a, b) then γ is simple, which means that it does not
intersect itself. Note that this does not include end points.
A contour, also called a path, is a finite sequence of directed smooth curves. Contours can
be open or closed; we will restrict to simple contours in what follows. When we integrate complex
functions, we usually do so over contours, not just real intervals.
Since a closed path starts where it finishes we need to also define its orientation. By convention
we take the integration in the anticlockwise direction.
(See Figure 7 on the next page.)
Integration along a contour The integral of a function f (z) along a curve γ is given by
Z Z b
f (z) dz = f (z(t)) z 0 (t) dt, (57)
γ a
where the curve is parameterised as z(t) with t real. The value of the integral does not depend on
the choice of this parameter t but only on the curve γ. (Proof by change of variable from t to τ , for
an arbitrary monotonic function t(τ ).) Note however that the direction of the integral is affected
by the orientation of the contour–this is why we need a convention (anticlockwise). To go in the
opposite direction would result in a change of sign.
Example: Integrate the function (z − a)n around a circle with radius r and center a, i.e., find
In = C(a:r) (z − a)n dz.
R
Cauchy’s Theorem Let f (z) be a function which is holomorphic in a region D containing the
contour γ. Then
Z
f (z) dz = 0. (62)
γ
This is the single most important theorem in complex variable theory. It follows directly from the
Cauchy-Riemann equations, and it can be proved in a similar way to the proof of Stokes’ theorem
in vector calculus.
13
Curve Path
Im Im
z(a)
γ
α β Re Re
z(b)
Re Re
Im Not closed
Contour
Re
14
Im
Re
Cauchy’s Residue Theorem Cauchy’s residue theorem tells us how to integrate functions with
isolated singularities over closed contours.
Let f be holomorphic inside and on a contour γ except for a finite number of poles {ak } inside
(not on) γ. Then Z X
f dz = 2πi Res(f : ak ). (63)
γ
singularities ak
of f (z)
inside γ
Proof outline: Cauchy’s residue theorem follows from Cauchy’s theorem and (61)—see fig-
ure 9. Each isolated singularity ak can be covered by a disk Dk with boundary C(ak : rk ), with
a sufficiently small radius rk so that no disk overlaps another. This implies that ak is the only
singularity inside Dk .
Inside this disk f (z) has Laurent series
∞
X
f (z) = cn (z − ak )n , (64)
n=−∞
so that Z Z ∞
X
f (z) dz = cn (z − ak )n dz
C(ak :rk ) C(ak :rk ) n=−∞
∞
X Z
= cn (z − ak )n dz
C(ak :rk )
n=−∞ (65)
∞
X 2πi, n = −1
= cn
0, n 6= −1
n=−∞
15
Im
a1 γ
a3
a2
Re
Here the up and down contours cancel each other out, since they are the same integral but in
opposite directions, so that γ ∗ only consists of:
γ going anticlockwise, with all the circles C(ak , rk ) going clockwise.
Hence Z Z XZ
0= f (z) dz = + f (z) dz − f (z) dz (67)
γ∗ γ k C(ak ,rk )
So Z X
f (z) dz = 2πi Res(f : ak ). (68)
γ k
The deformation theorem states that the integral around a closed contour is unchanged as
the contour is deformed, provided that the singularities contained within the contour remain the
same. One can therefore deform the integration contour without changing the value of the integral
providing that the deformation does not cross any singularity.
This is proved by applying the residue theorem to the function g(z)/(z − a), which by assumption
has a simple pole at z = a.
Example: Evaluate the integral f (z) = (z−1)12 (z−2) about the following contours:
(i) C(0 : 12 ); (ii) C(0 : 3); (iii) C(1 : 12 ); (iv) C(2 : 3).
16
Im
Im
Re
Re
Cauchy’s Theorem is independent of the contour, The Residue Theorem is independent of the contour
provided it remains inside an analytic region. provided the singularities inside remain unchanged.
Im
γ
iv
γ
ii
γ γ
i iii
1 2 Re
1
Figure 11: Sketch diagram for f (z) = (z−1)2 (z−2) .
(z − 2) 1
Res(f : 2) = lim (z − 2)f (z) = lim = = (2 − 1)−2 = 1. (70)
z→2 z→2 (z − 2)(z − 1)2 (z − 1)2 z=2
1 d d
Res(f : 1) = lim (z − 1)2 f (z) = lim (z − 2)−1 = −(z − 2)−2 z=1
= −1. (71)
(2 − 1)! z→1 dz z→1 dz
Alternatively we could look at the Laurent Series of f about each singularity to find the coefficient
17
1
of (z−a) : e.g, for z = 1
1 1 1 −1 −1
f (z) = (z−1)2 (z−2) = (z−1)2 −1+(z−1) = (z−1)2 (1 − (z − 1))
−1 1 1 1
= (z−1)2 1 + (z − 1) + (z − 1)2 + . . . = − (z−1) 2 − (z−1) −1− (z−1) − . . . ⇒ c−1 = −1.
Z (72)
1
(i) f (z) has no singularities inside C(0 : )⇒ f (z)dz = 0.
2 C(0: 21 )
(ii) f (z) has the singularities z = 1 and z = 2 inside C(0 : 3) ⇒
Z
f (z)dz = 2πi (Res(f : 1) + Res(f : 2)) = 2πi (1 − 1) = 0. (73)
C(0:3)
1
Example: Find the integral of 1−z 4 about each of the contours γ0 , γ1 , γ2 , γ4 , γ5 in figure 12.
Im 3i
2i
γ2 γ
4
i
γ1
−3 −2 −1 γ0 1 2 Re
−i
γ
5
−2i
1
Figure 12: Sketch diagram for f (z) = 1−z 4 .
1 4
f (z) = 1−z 4 has singularities when z = 1: it has four simple poles at z = 1, −1, i, −i. This
may be shown by considering
z 4 = 1 = ei0 = ei(0+2πk) , k ∈ Z. (76)
This gives the roots as:
41
zk = ei2πk = ei2πk/4 = eiπk/2 , k = 0, 1, 2, 3. (77)
18
Using additional values of k simply duplicates these roots; a fourth order polynomial has exactly
four roots.
Hence the roots are eiπ0/2 = e0 = 1, eiπ/2 = i, eiπ = −1, eiπ3/2 = e−iπ/2 = −i.
Using the “g/h0 ” rule we then get
1 1 1 zk zk
Res(f : zk ) = Res : zk = = − zk−3 = − 4 = − . (78)
1 − z4 −4z 3 z=zk 4 4zk 4
1
Example: Evaluate the integral of f (z) = z2 (1+sin z) about C(0 : 1); to find any residues first
derive the Laurent Series.
The function has simple poles when sin z = −1, but these are not inside C(0 : 1).
f (z) has a double pole at z = 0. Expanding about z = 0:
z3 1 1 1 1 1
1 + sin z = 1 + z − + ... ⇒ = 1 − z + ... ⇒ 2 = 2 − + ... (80)
3! 1 + sin z z 1 + sin z z z
Z
⇒ c−1 = −1 ⇒ f (z)dz = 2πiRes(f : 0) = −2πi. (81)
C(0:1)
Principle of the argument Consider a function f (z) as z travels along a contour γ. The curve
traced by f (z) is called the image of γ under f and is denoted by f (γ).
γ z(a)
f(γ )
z(b)
Im
f[z(a)] f[z(b)]
Re
Figure 13: The path f (γ) traced by f (z) as z follows the contour γ.
We define the Winding Number W as the number of times f (z) winds around the origin
(zero) as z goes along γ.
If N is the number of zeros of f (z) inside γ, counted by their multiplicity
and P is the number of poles of f (z) inside γ, counted by their multiplicity then
W = N − P. (82)
19
This is the Principle of the Argument (or Argument Principle) and is used in Nyquist theory:
see section 7.
Proof outline: Consider f near a zero a of order n. Then f (z) = (z − a)n g(z) where g(z) is
analytic (and nonzero) at a. We have
Z 0 Z 0
!
1 f (z) 1 n g (z) 1
dz = + dz = (2πi n + 0) = n. (since g 0 /g is analytic)
2πi γ f (z) 2πi γ z − a g(z) 2πi
(83)
Therefore, each zero of f contributes its multiplicity to the integral.
Similarly, consider f (z) near a pole b of order p. Then f (z) = (z − b)−p h(z) where h(z) is
analytic (and nonzero) at b. We have
Z 0 0
!
−p
Z
1 f (z) 1 h (z) 1
dz = + dz = (−2πi p + 0) = −p. (since h is analytic)
2πi γ f (z) 2πi γ z − b h(z) 2πi
(84)
Therefore, each pole of f contributes its multiplicity to the integral, with a minus sign.
1
R f 0 (z)
The integral 2πi γ f (z)
dz is sometimes referred to as the zero and pole counting integral.
0
f (z) d
On the other hand, we note that f (z) = dz (ln f (z)) so that
Z 0 Z
1 f (z) 1 d
dz = (ln f ) dz (85)
2πi γ f (z) 2πi γ dz
Z Z
1 d 1 d 1
= [ln |f (z)|]dγ + [i arg(f (z))]dγ = 0 + i 2πW =(86)
W.
2πi γ dz 2πi γ dz 2πi
Every time f (z) goes round zero once, ln |f (z)|, the real part of ln f (z), remains unchanged while
its imaginary part, the argument of f (z), increases by 2π.
2.10 M L estimates
We will often need to show that the integral along a curve vanishes as the curve is taken out to
infinity. Consider a function f (z) which is bounded so that it satisfies
|f (z)| ≤ M (87)
and it is particularly simple to work out when our contours are combinations of straight lines and
circular arcs.
Then an upper bound on the absolute value of the integral is
Z
f dz ≤ M L. (89)
γ
If M L → 0 in some limit then so does the absolute value of the integral. Sometimes this estimate
is not sharp enough, that is, the integral vanishes but lim M L does not. Then we need a better
estimate than an M L estimate
1
Example: The integral of z2 +a 2 (a 6= 0) about C(0 : R) = {z : |z| = R}.
20
2 2 i2θ
Im a+R e
2 2 i2θ
a+R e
a2 2
R
|a| 2
2 2 Re
R−|a|
If we let the circle grow to cover the entire complex plane, R → ∞, then
Z
2πR
→ 0, so that lim f (z) dz = 0. (91)
R2 − |a|2 R→∞ γ
R
(Both results are expected from the Residue Theorem: f (z) has two simple poles at ±ia, with
corresponding residues ±1/(2ia). For R a the contour C(0 : R) contains both poles, and since
their residues are equal and opposite in sign, the integral vanishes. For R a the contour does
not contain any singularity and the integral vanishes by Cauchy’s Theorem.)
To work out the maximum modulus of the integrand it is often useful to apply the triangle
inequalities. For any complex numbers z1 and z2 ,
|z1 + z2 | ≤ |z1 | + |z2 | (93)
while for any complex numbers z3 and z4
|z3 − z4 | ≥ ||z3 | − |z4 ||. (94)
(In the latter equality the overall modulus takes the absolute value, as |z3 | − |z4 | can be negative.)
Consider now
(z1 + z2 )
. (95)
(z3 + z4 )
The modulus of this expression is maximised when the modulus of the numerator is maximised
and the modulus of the denominator is minimised. Therefore, applying the triangle inequalities,
we obtain
(z1 + z2 ) |z1 + z2 | |z1 | + |z2 |
= ≤ , (96)
(z3 + z4 ) |z3 + z4 | ||z3 | − |z4 ||
where we used | − z4 | = |z4 |.
For example, if
1
f (z) = , (97)
z4 +α
with α an arbitrary complex number
1
|f (z)| ≤ . (98)
||z 4 | − |α||
21
Small circular arcs Imagine that the function f (z) which you are considering has a simple
pole a sitting on the contour you would like to integrate around. One way around this is to
circumnavigate the point a with a small circular arc—see Figure 15. (In practice we are usually
interested in semi-circular arcs or circles.)
Im
ε −> 0
Re
Our arc with radius is given by γ = {z(θ) = a + eiθ : θ1 < θ < θ2 } and should be orientated
anticlockwise. Hence Z Z θ2
f a + eiθ ieiθ dθ.
f (z) dz = (99)
γ θ1
Res(f : a)
f (z) = g(z) + (100)
z−a
where g(z) is analytic. Then
Z Z θ2
iθ Res(f : a)
f (z) dz = g(a + e ) + ieiθ dθ
γ θ1 a + eiθ − a
Z θ2 Z θ2
= g(a + eiθ ) ieiθ dθ + Res(f : a) idθ (101)
θ1 θ1
Z θ2
= g(a + eiθ ) ieiθ dθ + i Res(f : a) (θ2 − θ1 ) .
θ1
If γ = C(0 : ), given by θ1 = 0, θ2 = 2π, then this reduces to the Residue Theorem.
If we circumnavigate the singularity with a semicircular arc
22
2.11 Multifunctions
The archetype for all other multi-valued functions is the multi-valued function ln z, and so we
consider it as a prototypical example. Recall that
ln z = ln r + iθ
(105)
= ln |z| + i arg(z)
Im
z Log z = ln r + iθ
= ln r + i (θ+2π)
θ
Re
If we move z around the origin on a closed curve in an anticlockwise direction, coming back to
its starting point, then the value of z is unchanged. Re ln z = ln |z| also goes back to its starting
value, but Im ln z = arg z has increased by 2π. Therefore ln z is multi-valued—it is not a proper
function.
Branch points Note that ln z is not defined at zero, because ln |z| is not. The change in Im ln z
depends only on how often one goes around z = 0, not on the shape of the curve. The point z = 0
is called a branch point of ln z.
The function z a can be written as
As z goes around zero, z a keeps its modulus, but changes its argument by 2aπ. Therefore z = 0 is
a branch point of z a unless a is an integer. If a = 1/n, where n is an integer, then z a goes back to
what it was after one has gone around zero n times. z 1/n is called n-valued; one also says it has
n branches. For irrational a, z a has infinitely many branches; so does ln z.
For example, consider z 1/2 . This can be written as
i i
z 1/2 = |z|1/2 e 2 arg z = r1/2 e 2 θ , (107)
i.e. the multifunction picks up a minus sign. This multifunction has two branches, since it returns
to its original value after winding twice around zero.
One can make more complicated multi-valued functions from the multifunctions given above.
For example: √ √
z m/n is√n-valued;
√ sin z is double-valued; cos z is single valued (when z goes once around the
origin, z → − z, but cos z is even.)
23
Branch cuts To make f (z) single valued we can introduce branch cuts into the complex plane.
It connects a branch point either to infinity, or to another branch point.
As long as the path z takes (i.e. the contour it follows) does not cross a branch cut, the function
f (z) is single valued. If we cross the branch cut in any way, the value of the function jumps. The
position of the branch cut is arbitrary, as long as it ends at the branch point (if we think of infinity
as one branch point).
Im
branch point
Re
cut
jump
Going to infinity in any direction can be thought of as going to a single “point at infinity”.
This point is a branch point of, for example, ln z in the precise sense that zero is a branch point of
ln z1 . If we turn infinity into a point, then any branch cut links two branch points. It is sometimes
useful to make use of the extended complex plane, defined as C ∪ ∞.
ln z has a branch cut going from zero to infinity. One can go to “infinity” in any direction, for
example to ∞, to −i∞, p or to (3 − i)∞.
Example: f (z) = z(z − 1)
f (z) has the two branch points 0 and 1. f (z) is double-valued around either of them.
There exist two possible types of branch cuts: One linking 0 to 1, and the other linking both
to infinity (see figure 18). Which one we choose will depend on what problem we are considering.
Im Im Im
0 1 Re 0 1 Re 0 1 Re
p
Figure 18: Possible branch cuts (the dashed line) for f (z) = z(z − 1).
Branch cuts define the domain of a function. For ln z, the introduction of a branch cuts means
24
we restrict the values that θ can take. For example:
a branch cut from 0 to +∞ gives the restriction 0 ≤ θ < 2π;
a branch cut from 0 to −∞ gives the restriction −π < θ ≤ π.
Re Re
1 1 1 θ
z 2 = rei(θ+2πn) 2 = r 2 ei( 2 +πn) (where n ∈ Z) gives two distinct values, namely
1 θ 1 θ
r 2 ei 2 and r 2 ei( 2 +π) = −r1/2 eiθ/2 , ⇒ f (z) is a multifunction.
f (z) has a branch point at z = 0.
f (z) has simple poles at z = 2i and z = −2i.
On γR , z may be parameterized by z(θ) = Reiθ (0 ≤ θ < π), so dz = iReiθ dθ.
Z Z π √ iθ
Re iθ
f (z) dz = 2 iRe dθ. (109)
γR 0 4 + (Reiθ )
Using M L estimates:
√
Reiθ R2
1
Letting the semicircle cover the entire upper half of the complex plane:
Z 3
πR 2
f (z) dz < → 0, R → +∞. (111)
γR R2 − 4
Hence Z
lim f (z) dz = 0. (112)
R→∞ γR
25
Im
+R i
γR
2i
γε
γ2 γ1
Re
−R −ε ε +R
−2i
Hence Z
lim f (z) dz = 0. (114)
→0 γ
The line γ1 may be parameterized by z = x, for
< x < R. Hence Z Z R 1
x2
f (z) dz = dx. (115)
γ1 4 + x2
As → 0 and R → ∞ this tends to the integral
Z ∞ 1
x2
I = dx. (116)
0 4 + x2
The line γ2 may be parameterized by z = x, for −R < x < −. We must be careful now as x
is negative:
Z Z − 1 Z −R 1 Z R 1 Z R 1
x2 x2 (−X) 2 X2
f (z) dz = 2
dx = − dx = − − dX = i dX.
γ2 −R 4 + x − 4 + x2 4 + (−X)2 4 + X2
(117)
Here in the last two equalities we used the redefinition X = −x with X positive. As → 0 and
R → ∞ the last equality tends to the integral
Z ∞ 1
X2
i dX = iI, (118)
0 4 + X2
where I was the original integral of interest. (Remember that you can always relabel an integration
variable!)
If we consider the entire contour γ to be the union of γR , γ2 , γ , γ1 in the combined limit
→ 0, R → ∞ then the integral along γ becomes
Z Z Z Z Z
= limR→∞ lim→0 + + +
γ γR γ γ1 γ2 (119)
= 0 +0 +I +iI = (1 + i)I.
26
On the other hand, we have the Residue Theorem, which tells us that:
Z X
f (z) dz = 2πi Res(f : ak ) = 2πiRes(f : 2i)
γ
singularities ak
of f (z)
inside γ
1
z2 1 1 π 1 1−i π
= 2πi = πi z − 2= πi2− 2 e−i 4 = πi √ √ = (i + 1).
π
2z z=2ei 2 2 2 2
z=2i
(120)
To obtain these formulae, note that the residue in this case is most conveniently calculated with
what we called earlier the g/h0 formula. The function of interest is
1
z2
f (z) = 2
(121)
z +4
1
To calculate the residue, we identify g(z) = z 2 and h(z) = (z 2 + 4). Then h0 (z) = 2z giving the
residue used above. We have to be careful when we take the square root: the value of a square
root is most easily calculated by writing the complex number in polar form. In the case at hand,
we remember the standard relation that i = exp(iπ/2).
Equating (119) and (120), we finally obtain
Z ∞ 1
x2 π
I = 2
dx = . (122)
0 4 + x 2
Note that because I is a real integral our result must also be real.
Compare this to the result from real integration:
√
1
∞
Z ∞
√ √
x2 1 2−2 x+x 1 1
dx = ln √ + arctan x − 1 + arctan x + 1
0 4 + x2 4 2+2 x+x 2 2 0
1 1π 1π 1 1 π 1 π (123)
= ln 1 + + − ln 1 + − +
4 22 22 4 2 4 2 4
π
= − 0.
2
3.1 Semicircle contour: for integrals over the whole real line
Examples of functions can be integrated with a semicircle contour (see figure 21) include
Z ∞
P (x)
I = dx (124)
−∞ Q(x)
27
∞
eikx P (x)
Z
and dx, (125)
−∞ Q(x)
where P and Q are polynomials, and Q(x) has no zeroes on the real line. P/Q is called a rational
function.
The order of the polynomial Q must be at least the order of P plus two, in order for the large
semicircle contribution to vanish by an M L estimate.
Im Im
γ
R
−R R
γ
1
γ Re Re
1 −R R
γ
R
Figure 21: Closing a contour on the real line with a semicircular arc. The left diagram is appropriate
for k < 0 and the right diagram is appropriate for k > 0. Note that we transverse the real line in
different directions.
Whether we close the contour above or below depends on the value of the integrand on the
semicircle, i.e., where z = Reiθ , as R → ∞.
For integrals of the form (124) we can close the contour above or below.
If k > 0 in (125) then we must close the contour above because eikx is exponentially small in
the upper half plane, but exponentially large in the lower half plane. Conversely, if it is k < 0 we
must close in the lower half.
If P and Q are real polynomials, we can take the real and imaginary parts of this integral to
obtain cos kx and sin kx in the integrand instead of eikx . For example
Z ∞ Z ∞ eikx dx
cos kx dx
2
= Re 2
. (126)
−∞ 1 + x −∞ 1 + x
Z ∞
dx
Example: Evaluate I = 2 2
.
−∞ (1 + x )
1
I F (x) = (1+x 2 )2 is of the form (124) so we choose a semicircular contour, which we may close
above or below.
II The function f (z) = (1+z1 2 )2 is equal to F (x) on z = x.
f (z) has singularities at z = ±i , which do not lie on the contour provided R > 1.
III Consider the contour γ in two parts:
Z Z +R
1
F (z) dz = dx → I, R → ∞. (127)
γ1 −R (x2
+ 1)2
( )
π Z π
|R||eiθ |
Z Z
1 iθ
F (z) dz = 2 Re dθ < dθ max 2
γR 0 ((Reiθ ) + 1)2 0 0≤θ≤π |R2 ei2θ + 1|
(128)
πR
≤ → 0, R → ∞.
(R2 − 1)2
28
Im
γ
R
i
γ
1
Re
−R R
−i
29
we have an inverse FT, and so this exponential is eikx , but in a FT it would be e−ikx , so be
careful.) Since the sign of the complex exponential is determined by the sign of x (remember we
are integrating with respect to k) , we need to consider x > 0 and x < 0 separately.
Let us first take x > 0 so that we close above (see figure 23).
Im
γ
γ
R
ai
γ
1
−R R Re
−ai
because |eikx | = |eix Re(k) | |e−x Im(k) | = e−x Im(k) and Im(k) = Im(Reiθ ) = R sin θ > 0 for θ ∈ (0, π),
so that e−x R sin θ → 0, R → ∞ for x > 0.
This is why we had to specify the sign of x and (for x > 0) close in the upper half plane.
In the limit R → ∞ we have:
Z
1 1 X
f (x) = F (k) dk = 2πi Res(F : ak ) (137)
2π γ 2π
singularities ak
of F (k)
inside γ
30
For x < 0 we close the contour below so that it now contains the singularity −ia and not +ia.
Here we have minus f (x) on the left hand side because going around semicircular contour closed
below in an anticlockwise direction means we integrate F (k) from +R to −R.
−ax
e , x > 0,
Hence f (x) = ⇒ f (x) = e−a|x| . (140)
e+ax , x < 0,
γ
γ2
π/2 i
γ1
Re
R−>
R∞ dx
Figure 24: The box contour for I = −∞ cosh x
, which has poles at z = (2n + 1)πi/2 for all integers
n. Here we choose a = π.
we note that the real part vanishes for y = (2n + 1)π/2, with n an integer or zero, while the
imaginary part vanishes either at y = nπ or x = 0. However at y = nπ the real part does not
vanish, so the locations at which both real and imaginary parts vanish are x = 0, y = (2n + 1)π/2.
One can use M L estimates to show that the contributions from the vertical parts of the rectangle
vanish as the length of the rectangle is taken to be large. Clearly along γ1 we obtain
Z Z R
dz dx
= . (145)
γ1 cosh z −R cosh x
31
Along γ2 , z = x + iπ and hence
Thus Z Z −R Z R
dz dx dx
= = . (147)
γ2 cosh z R (− cosh x) −R cosh x
(Note that the integration on this part of the contour is in the direction of decreasing x.) Therefore
Z Z R
dz dx
= 2 LimR→∞ = 2I. (148)
γ cosh z −R cosh x
Now we can compute the integral over the closed contour using the residue theorem; since the
residue of the pole is −i Z
dz
= 2πi(−i) = 2π (149)
γ cosh z
and therefore I = π.
There are many more functions that can be integrated over the real line. In each case it is
important to verify that it is possible to obtain both the estimates (for those parts of the contour
not on the line, such as the semicircle with radius R) and residues (for singularities inside the
contour), since otherwise we will not obtain the (real) value of the integral we require.
3.5 Integrals over the half line: using a keyhole contour with multifunc-
tions
Although multifunctions complicate matters, this can be used to our advantage.
For a (single valued) function f (z) we have, for z = reiθ :
f (reiθ ) = f rei(θ+2π) . (151)
Consider two contour sections γ1 and γ2 , just above and just below the positive real axis (see
figure 25), that is, on either side of a branch cut for ln z. Then
32
We can therefore evaluate real integrals of the form
Z ∞
f (x)dx (156)
0
where γ is a “keyhole” contour—see figure 25. One must check, though, that both the large and
small circle give vanishing contributions as → 0 and R → ∞, and so it is best to derive this trick
from scratch rather than memorize it. Effectively, Eq. (157) allows us to calculate real integrals
R∞ R0
0
(or −∞ ) using the Residue Theorem—very convenient indeed!
Im
+R i
γR
γε
γ
1
Re
−R
ε γ +R
2
Figure 25: The keyhole contour for a branch cut [0, −∞): to cover the complex plane we let R → ∞
and → 0.
The keyhole contour (but without the ln z trick) also works for
Z ∞
xa f (x) dx (158)
0
a
where a is not an integer and
Z ∞ where f (z) is single-valued. This is because x is a multifunction.
dx
Example: Evaluate I = .
0 1 + x4
We will try the keyhole contour with
ln z
f (z) = . (159)
1 + z4
f (z) has a branch point at z = 0, so we use the branch cut z = x (x > 0).
π
f (z) has simple poles when z 4 = −1: these are given by ak = ei(2k+1) 4 for k = 0, 1, 2, 3. Explicitly:
π +1 + i 3π −1 + i 5π −1 − i 7π +1 − i
a0 = ei 4 = √ , a1 = ei 4 = √ , a2 = ei 4 = √ , a3 = ei 4 = √ . (160)
2 2 2 2
33
Im
+R i
γR
γε
γ
1
Re
−R +R
ε γ
2
Figure 26: By sketching our contour we note that we require <1 and R> 1 for all four singularities
to be inside the contour.
Next draw a keyhole contour γ as in figure 26, and consider its four pieces:
Z Z Z Z Z
f (z) dz = f (z) dz + f (z) dz + f (z) dz + f (z) dz
γ γ1 γR γ2 γ
Z ∞ Z 2π Z 0 Z 0
ln x ln R + iθ ln x + 2πi ln + iθ
= dx + iReiθ dθ + dx + ieiθ dθ.
0 1 + x4 0 1 + R4 ei4θ ∞ 1 + x4 2π 1 + 4 ei4θ
(161)
Now Z 2π
ln R + iθ ln R + 2π
iReiθ dθ < (2πR) → 0, R → ∞, (162)
0 1 + R4 ei4θ R4 − 1
and similarly Z 2π
ln + iθ 2π (ln + 2π)
− ieiθ dθ < → 0, → 0. (163)
0 1 + 4 ei4θ 4 − 1
Hence Z Z
ln z
f (z) dz = dz = −2πiI (164)
γ γ1 +γ2 1 + z4
[using Eq. 157]. But we also have
Z X
f (z) dz = 2πi Res(f : ak ). (165)
γ
singularities ak
inside γ
34
Evaluating all four residues:
ln z i
(2k+1)π
Res(f (z) : ak ) = Res : e 4
+ z4
1(2k+1)π
i
ln e 4
0
= (2k+1)π 3 using “g/h rule”
4 ei 4
i (2k+1)π
4 i
(2k+1)π
= (2k+1)π 4 e 4
4 ei 4 (167)
i (2k+1)π (2k+1)π
= 4
ei 4 (using a4k + 1 = 0)
−4
√ √ √ √
= i 1π 3π 5π 7π
16 (−1 − i) / 2 + i 16 (+1 − i) / 2 + i 16 (+1 + i) / 2 + i 16 (−1 + i) / 2
√ (168)
= iπ [(−1 + 3 + 5 − 7) + i(−1 − 3 + 5 + 7)] /16 2
√
= iπ [8i] / 16 2
π
I= √
2 2
(which is real as required).
where n is an integer and f (z) is single-valued. In this case we can choose the contour such that
the integrals along the two straight sections of the “pie” are the same, up to an overall factor. For
large n, this will be simpler than the keyhole contour because fewer residues need to be calculated.
Consider, for example, the integral
Z ∞
dx
. (170)
0 1 + x7
1
The singularities of 1+z 7 are plotted in figure 27. A segment contour can be drawn to include
only a single residue. Z ∞
1
Example: Evaluate (once again) I = dx.
0 1 + x4
1
Consider f (z) = 1+z 4 ,
which has four singularities, one in each quadrant.
We will use a segment contour covering only the first quadrant—see figure 28.
35
R∞ dx
Figure 27: The segment contour for 0 1+x7 .
γ
γ2
γR
2−1/2i
γ1
2−1/2 2−1/2
−1/2
−2 i
R∞ dx
Figure 28: The segment contour for 0 1+x4 .
On γ1 , z = x, 0 < x < R:
Z Z R
f (z) dz = f (x) dx → I, R → ∞. (171)
γ1 0
36
π 1+i
The only singularity of f (z) inside the contour γ is a0 = ei 4 = √ .
2
1 a0 a0 1+i
Res(f : a0 ) = 3 = 4 =− =− √ (174)
4a0 4a0 4 4 2
Note that we cannot use the residue calculated in (167) as this was for a different function, namely
1
1+z 4 ln z. Although we are considering the same real integral we are using a different contour with
a different integrand.
Using the Residue Theorem we now have
−1 − i
Z
π(−i + 1)
I(1 − i) = f (z) dz = 2πiRes(f (z) : a0 ) = 2πi √ = √ (175)
γ 4 2 2 2
π
giving I = √
2 2
as before.
Im
Re
R∞ x1/2 dx
Figure 29: The keyhole segment contour for 0 1+x6 . The dashed line is the branch cut.
Therefore the integral over the closed contour is equal to the contributions from the two straight
lines: Z R 1/2 Z
z 1/2 (xeiα )1/2
Z
x
6
dz = dx + d(xeiα ), (178)
γ 1+z 1 + x6 R 1 + (xe )
iα 6
where α is the angle of the segment. If we choose 6α = 2π, namely α = π/3, then the second term
becomes Z 1/2 Z R 1/2
iπ/2 x x
e 6
= −i (179)
R 1 + x 1 + x6
and is therefore proportional to the integral which we wish to compute. Computing the contour
integral with the residue theorem as usual, we find that
z 1/2
Z
π
6
dz = √ (1 − i) (180)
γ 1 + z 3 2
37
and therefore
∞
x1/2
Z
π
= √ . (181)
0 1 + x6 3 2
A semicircle is, of course, just a special case of a segment, and a keyhole semicircle contour
may sometimes be appropriate. In all of the above examples one can do well with just the semicircle
and keyhole contours; however, segment/box contours may be easier and more efficient, as they
enclose fewer singularities.
(recall tan has period π, not 2π) we use as a contour the unit circle:
Remember that we need only consider singularities inside the unit disk.
Im
z=exp( iθ)
θ =0 Re
θ =2π 1
Figure 30: The unit circle contour for periodic functions: z = eiθ ⇒ dz = ieiθ .
It follows from
1
z = eiθ = cos θ + i sin θ, = e−iθ = cos θ − i sin θ, (184)
z
that
1 1
cos θ = (z + z −1 ), sin θ = (z − z −1 ), (185)
2 2i
and similarly cos 2θ = 21 (z 2 + z −2 ), etc.
If f is a rational function
Z 2πof sin θ and cos θ, then it is also a rational function of z.
dθ
Example: Evaluate I = .
0 5 − 4 cos θ
The integrand is a periodic function so we use the contour γ = C(0 : 1).
The choice of f (z) is given by transforming θ → z:
1 1 dz
cos θ = z+ , dθ = (186)
2 z iz
38
dθ 1 dz −1 dz 1
= = so use f (z) = . (187)
4 1 2i z − 52 z + 1 − 52 z + 1
5 − 4 cos θ 5− 2 z + z iz 2 z2
Then: Z Z Z 2π
dz dθ
f (z) dz = = −2i = −2iI. (188)
γ C(0:1) z 2 − 52 z + 1 0 5 − 4 cos θ
Z
Evaluate f (z) dz using the Residue Theorem.
γ
5 1
z 2 − z + 1 = (z − 2) z− (189)
2 2
Im
1/2 1 2
Re
1
Figure 31: The unit circle contour for f (z) = z 2 − 25 z+1
.
1 1
R
γ
f (z) dz = 2πi Res f : 2 (since only 2 inside γ)
= 2πi 1 (190)
z−2
z= 12
1
= 2πi − 32
= − 4πi
3
4π
Equating we have −2iI = 3 i which gives I = + 2π
3 .
4 Fourier transforms
4.1 Derivation of the Fourier transform from the complex Fourier series
You have already covered the complex Fourier series in Year 2. If f (x) is a periodic function with
period L, that is f (x + L) = f (x), then we can write it as
∞
X 2πn
f (x) = cn eikn x , kn := , (191)
n=−∞
L
39
where the complex Fourier coefficients cn are given by
Z L2
1
cn = f (x)e−ikn x dx. (192)
L − L2
Note the integral could be taken over any other full period, for example from 0 to L. If f (x) is
real, then cn = c∗−n , where the star denotes the complex conjugate.
[You can verify (192) directly by changing n to m and then substituting the expression (191)
for f (x). After integrating, you get cm = cn , as required.]
From the complex Fourier series we can obtain the Fourier transform as a limit. Define
2π
fˆ(kn ) := cn L, ∆k := kn+1 − kn = . (193)
L
With this notation, (191) and (192) become
∞
1 X ˆ
f (x) = f (kn )eikn x ∆k, (194)
2π n=−∞
Z L2
fˆ(kn ) = f (x)e−ikn x dx. (195)
−L
2
So far, the function fˆ(k) is defined only at the discrete points kn . Now as weP
take the limit
R L → ∞,
we have ∆k → 0 and we can think of it as dk under an integral, that is ∆k → dk. At the
same time, fˆ(k) is then defined on the continuous real line. We have
Z ∞
1
f (x) = fˆ(k)eikx dk, (196)
2π −∞
Z ∞
fˆ(k) = f (x)e−ikx dx. (197)
−∞
(Note that in MATH3084/6162 I have a different convention from the one in MATH3083/6163.)
When f (x) is real, then fˆ(k) = fˆ(−k)∗ (the complex conjugate).
40
Properties Fourier transforms have the following basic properties:
F [af (x) + bg(x)] = afˆ(k) + bĝ(k) (201)
F [f (x − a)] = e−ika fˆ(k) (202)
1 ˆ
F [f (ax)] = f (k/a) (203)
|a|
d
F [xf (x)] = i fˆ(k) (204)
h i dk
F f (m) (x) = (ik)m fˆ(k) (205)
Z ∞
F [f ∗ g] := F f (y)g(x − y)dy = fˆ(k)ĝ(k) (206)
−∞
Most of these properties place conditions upon the function f (x). These conditions are usually
satisfied if f (x) and its derivatives are integrable and differentiable.
The Plancherel theorem (sometimes also called the Parseval theorem) says that if f (x) is
square-integrable, then so is fˆ(k), and the two obey
Z ∞ Z ∞
2
|f (x)| dx = |fˆ(k)|2 dk. (207)
−∞ −∞
4.3 Proofs
The shifting theorem (202) can be proved by a change of variable:
Z ∞ Z ∞
F [f (x − a)] = f (x − a)e−ikx dx = f (X)e−ik(X+a) dX = e−ika fˆ(k), (208)
−∞ −∞
where we set X = x − a. For the product theorem (203) we again change variable:
Z ∞ Z ∞
−ikx dX 1ˆ
F [f (ax)] = f (ax)e dx = f (X)e−ikX/a = f (k/a) (209)
−∞ −∞ a a
where now X = ax and we assume a is positive. (If a is negative we obtain a modulus sign as
given in (203).)
Now consider (204)
Z ∞ Z ∞
dfˆ(k)
d
F [xf (x)] = xf (x)e−ikx dx = i f (x)e−ikx dx = i , (210)
−∞ dk −∞ dk
where we have assumed that the integration and differentiation commute.
The transform of derivatives (205) is proved using integration by parts:
h i Z ∞ h i∞ Z ∞
F f (m) (x) = f (m) (x)e−ikx dx = f (m−1) (x)e−ikx − (−ik)f (m−1) (x)e−ikx dx,
−∞ −∞ −∞
(211)
and the boundary term vanishes provided that f (m−1) (x) vanishes as x → ±∞. Integrating by
parts repeatedly we thence obtain
h i Z ∞
(m)
F f (x) = (ik) m
f (x)e−ikx dx. (212)
−∞
The convolution theorem is proven as for the Laplace convolution theorem, see Sec. 5.1 below.
41
This is another way of saying that the Fourier transform is its own inverse, up to a factor of 2π
and changing one sign.
For even or odd functions, the Fourier transform is its own inverse, up to a factor of 2π and a
sign. It can be shown that
h i
Even: if f (x)= f (−x) then F [F[f (x)]] = F fˆ(k) = 2πf (x)
h i (214)
Odd: if f (x)=−f (−x) then F [F[f (x)]] = F fˆ(k) = −2πf (x)
Fourier transforming an even function twice brings it back to 2π times that function,
Fourier transforming an odd function twice brings it back to −2π times that function.
δ(x − y) = 0 x 6= y (215)
and Z ∞
f (y)δ(x − y)dy = f (x) (216)
−∞
(Strictly speaking, the Dirac delta function is not a function: it is a what is called a distribution,
also called a generalised function. However it can be defined as the limit of a function and was
introduced by Cauchy and Poisson in this way.)
From the definition of the inverse Fourier transform
Z ∞
1
f (x) = fˆ(k)eikx dk (218)
2π −∞
we deduce that Z ∞ Z ∞
1 −iky
f (x) = f (y)e dy eikx dk (219)
2π −∞ −∞
and hence Z ∞ Z ∞
1
f (x) = f (y)eik(x−y) dkdy (220)
2π −∞ −∞
we obtain Z ∞
f (x) = f (y)δ(x − y)dy, (222)
−∞
implying that the Fourier transform of a constant function is proportional to the Dirac delta
function.
42
4.6 The δ-function as a limit of sharply peaked functions
Consider any integrable function h(x) whose integral over the real line is one, that is
Z ∞ Z ∞
h(x) dx = h(x)e−i·0·x dx = ĥ(0) = 1. (224)
−∞ −∞
and correspondingly
Z ∞ Z ∞
1
f (x, y) = fˆ(k1 , k2 ) exp[i(k1 x + k2 y)]dk1 dk2 . (233)
(2π)2 −∞ −∞
More generally, letting x denote a vector with n Cartesian components and k similarly denote an
n-dimensional vector Z ∞ Z ∞
fˆ(k) = ··· f (x)e−ik·x dx1 · · · dxn (234)
−∞ −∞
with Z ∞ Z ∞
1
f (x) = ··· fˆ(k)eik·x dk1 · · · dkn . (235)
(2π)n −∞ −∞
43
4.8 Application of Fourier transforms to partial differential equations
Partial Differential Equations (PDEs) concern functions of more than one variable.
In many practical examples we need to consider a function u(x, t) where x ∈ (a, b) and t ≥ 0,
are time and space (in one dimension). For example,
function u(x,t)
the Fourier transform fˆ(k, t) of f with respect to x has the following property:
This results can be extended for higher order derivatives, with each higher order derivative in x
requiring an additional condition, e.g.,
F [ft (x, t)] = fˆt (k, t), F [ftt (x, t)] = fˆtt (k, t). (241)
In many cases a second order PDE in f (x, t) can be transformed to a second order ODE in
ˆ
f (k, t), which can usually be solved more easily than the original PDE. The latter is an ODE in
the sense that only t derivatives but no x or k derivatives remain.
44
Example: Solve ut = ux − u with BC u → 0 as |x| → ∞ and IC u(x, 0) = 1/(1 + x2 ); you may
use the result
F 1/(1 + x2 ) = πe−|k| .
(242)
The modulus sign in the integrand suggests we should consider k > 0 and k < 0 separately:
Z ∞ Z 0
2et u(x, t) = e−k ei(t+x)k dk + ek ei(t+x)k dk
0 −∞
Z ∞ Z 0
= e(−1+i(t+x))k dk + e(+1+i(t+x))k dk
0 −∞
∞ 0
e(−1+i(t+x))k e(1+i(t+x))k
= +
−1 + i(t + x) 0 1 + i(t + x) −∞ (245)
1 1
= 0 − −1+i(t+x) + 1+i(t+x) − 0
−2 2
= = .
i2 (t + x)2 − 1 1 + (t + x)2
Hence u(x, t) = e−t /(1 + (t + x)2 ).
Example: Invert πe−|k| e(ik−1)t by inspection.
πe−|k| = F 1+x 1
2 using (242)
πe−|k| eitk = F 1
1+(x+t)2 using (202) (246)
e−t
πe−|k| eitk e−t = F 1+(x+t)2 e−t independent of k
1
Example: Solve the wave equation c2 ftt = fxx with BCs f, fx → 0 as |x| → ∞ and ICs
f (x, 0) = g(x), ft (x, 0) = 0.
45
Solve and impose ICs:
fˆ = A(k) cos ckt + B(k) sin ckt (249)
−ckA sin 0 + ckB cos 0 = 0 ⇒ B(k) = 0, and then ĝ(k) = A(k) cos 0 ⇒ A(k) = ĝ(k). (250)
So:
fˆ(k, t) = ĝ(k) cos(ckt). (251)
1 ickt
+ e−ickt , so we may write
Now cos(ckt) = e
2
1
fˆ(k, t) = ĝ(k) eictk + ĝ(k) e−ictk .
(252)
2
Using (202) we then get
1
(g(x + ct) + g(x − ct) ) .
f (x, t) = (253)
2
In the examples above, Fourier transforms were applied to partial differential equations, with
the transforms being taken along infinite spatial directions.
These transforms are unique, so that either one contains all the information necessary to reconstruct
f (x) when inverted.
It can easily be shown from the inversion formula for the complex Fourier transform that
2 ∞ ˆ 2 ∞ ˆ
Z Z
f (x) = fc (k) cos kx dk = fs (k) sin kx dk. (256)
π 0 π 0
π
Both the sine and cosine
h transforms
i are their own inverse, up to a factor of 2.
x
Example: Find Fs 1+x4 .
x
We let f (x) = 1+x4 and note that f (x) is an odd function, so that (letting X = −x)
∞ −∞ 0
−X
Z Z Z
x X
fˆs (k) = sin kx dx = (− sin kX) − dX = (sin kX) dX.
0 1 + x4 0 1 + X4 −∞ 1 + X4
(257)
Combining gives Z ∞
1 x
fˆs (k) = sin kx dx. (258)
2 −∞ 1 + x4
Using eikx = cos kx + i sin kx gives
Z ∞
1 x
fˆs (k) = Im ikx
e dx . (259)
2 −∞ 1 + x4
We can solve this using a semicircular contour as before, closed in the upper half plane for k > 0
and in the lower half plane for k < 0.
On γR z = Reiθ , θ ∈ (0, π) so
Z
R
F (z) dz ≤ (π R) → 0, R → ∞. (260)
γR R4 − 1
46
Im
γ
R
γ
1
−R R Re
z
Figure 33: The integral contour for F (z) = 1+z 4 eikz for k > 0.
On γ1 z = x, x ∈ (−R, R) so
Z Z +R Z ∞
x x
F (z) dz = eikx dx → eikx dx, R → ∞. (261)
γ1 −R 1 + x4 −∞ 1 + x4
π − √k k
= e 2 sin √ (k > 0)
2 2
47
For k < 0 we similarly obtain
π √ −k k
fˆs (k) = − e 2 sin √ (k < 0) (267)
2 2
Hence,
π −√
|k| k
fˆs (k) = e 2 sin √ . (268)
2 2
f(x) f(x)
odd even
x x
Figure 34: Odd and even extensions of a function defined on the half line.
Z ∞ Z −∞
= f (x) e−ikx dx − f (−x) e−ikx dx
0 0
Z ∞ Z ∞
−ikx
= f (x) e dx + f (X) e+ikX dX (271)
0 0
∞
eikx + e−ikx
Z
= 2 f (x) dx
0 2
Z ∞
= 2 f (x) cos kxdx
0
= 2Fc [f (x)] .
48
Similarly, the Fourier transform of an odd extension is given by
fˆodd (k) = −2i fˆs (k); (note the additional factor −i). (272)
Our choice of sine or cosine transform will depend upon the boundary conditions
of the PDE problem we want to solve—more on that in the next section.
The multiplication with −k 2 is just as in the complex Fourier transform, but now there are also
boundary terms reminiscent of those in the Laplace transform.
x x x
or
so that sine transform of fx brings in the cosine transform of f and vice versa; this is of little use
as fˆc (k, t) and fˆs (k, t) are fundamentally different functions of k.
49
Example: Use the Fourier cosine transform to solve uxx = c ut (c > 0) for x > 0, t > 0, with
−x2
u(x, 0) = e 2 and where ux (0, t) = 0; u(x, t) → 0 as x → ∞; and ux (x, t) → 0 as x → ∞.
You may use the result h x2 i √ k2
F e− 2 = 2πe− 2 . (276)
1 ˆ
By using property (203), namely that F [f (a x)] = |a| f (k/a) we can easily derive the more
general result r
√ 2
h
−α x2
2 i
−
( α x) 2π − 1 k2
F e = F e 2 = e α 2 , α > 0. (277)
α
√
Our result is restricted to α > 0 because here α = a ∈ R.
We write û = Fc [u], so that
−k 2 û − ux (0, t) = c ût (using BCs at ∞)
(278)
−k 2 û = c ût (using BC at 0)
2 k2
ût = − kc û ⇒ û(k, t) = û(k, 0) e− c t .
Z ∞
û(k, 0) = u(x, 0) cos kx dx
0
Z ∞
1 x2
= Re e− 2 e−ikx dx
2 −∞
1 h x2 i
= Re F e− 2 (279)
2
1 √ k2
= Re 2πe− 2 [Using Eq. (277) with α = 1]
2
r
π − k2
= e 2
2
r r
π − k2 2
− kc t π −(1+2 t ) k2 .
Hence û(k, t) = e 2 e = e c 2
2 2
Z ∞r
2 π −(1+2 t ) k2
u(x, t) = e c 2 cos kx dk
π 0 2
∞
r Z
1 2 t k2
= Re e−(1+2 c ) 2 e−ikx dk
2 π −∞
1 ht k2
i
= √ Re F e−(1+2 c ) 2 (but from k to x) (280)
2π
s !
x2
1 2π − 1 t 2 t
= √ Re e 1+2
c using Eq. (277) with α = 1 + 2
2π 1 + 2 ct c
c x2
q
= c
c+2t e− 2(c+2t) .
50
The Fourier sine and cosine transforms are applicable to problems on semi-infinite domains.
For example, recall the heat conduction equation
Tt = κTxx . (282)
Suppose this is solved subject to T (0, t) = T0 and T (∞, t) → 0; such boundary conditions imply
that the Fourier sine transform is appropriate, i.e. the transformed equation is
Another common application of Fourier sine and cosine transforms is to potential theory, i.e.
solutions to the Laplace equation in two dimensions
on a semi-infinite domain x > 0 and y > 0. Consider the following boundary conditions for this
equation: φ = 1 on x = 0 and y > 0; φ = 0 on y = 0 and x > 0; ∇φ → 0 as x, y → ∞.
Applying the Fourier sine transform in the x direction and using the boundary condition at
x = 0, we obtain
∂y2 φ̂s (k, y) − k 2 φ̂s (k, y) + k = 0. (285)
This is an inhomogeneous second order equation, for which the general solution is
1
φ̂s (k, y) = + Aeky + Be−ky . (286)
k
Since φ̂s (k, y) → 0 as y → ∞, the integration constant A must vanish. Since φ̂s (k, 0) = 0
1
0= +B (287)
k
and hence
1
φ̂s (k, y) = (1 − e−ky ) (288)
k
which in turn implies that
Z ∞
2 1 2 y
φ(x, y) = (1 − e−ky ) sin(kx)dk = tan−1 . (289)
π 0 k π x
(Note: you would not be expected to know integral identities such as this one, but would be given
the required identity on a coursework or exam question!)
5 Laplace transforms
The Laplace transform transforms a real function f (t) of a real variable t ≥ 0 (usually t represents
time) to a complex function f¯(s) of a complex variable s.
The Laplace transform is defined by
Z ∞
¯
L [f (t)] = f (s) = f (t)e−st dt. (290)
0
Again we note that the convention of placing a bar above f to denote its transform is not universal:
some books use F for the transform of f , others the reverse.
If f (t) is piecewise continuous for t ≥ 0 and there exist constants M, α > 0 such that
for all t > T for some T > 0, then L[f (t)] exists and is analytic in the half-plane Re(s) > α.
One can derive the Laplace transform from the Fourier transform, but each has its own domain
of usefulness. Note, for example, that a function f (t) that grows exponentially as eαt still has a
Laplace transform, but that it would not have a Fourier transform.
51
Im 1111111
0000000
0000000
1111111
s−plane
0000000
1111111
0000000
1111111
0000000
1111111
0000000
1111111
0000000
1111111
0000000
1111111
0000000
1111111
0000000
1111111
0000000
1111111
α 1111111
0000000
0000000
1111111
Re
0000000
1111111
0000000
1111111
0000000
1111111
0000000
1111111
0000000
1111111
0000000
1111111
0000000
1111111
0000000
1111111
Figure 36: The half plane Re(s) > α. All the singularities of f¯(s) lie to the left of Res(s) = α.
Z ∞
L [sin at] = f¯(s) = sin at e−s t dt
0
∞ Z ∞
e−s t
−a
= sin at − cos at e−s t dt (integration by parts)
−s 0 0 s
Z ∞
= 0+ a
s cos at e−s t dt
0
∞ Z ∞
e−s t e−s t
a
= s cos at − −a sin at dt (by parts again)
−s 0 0 −s
Z ∞
a
= a
s 0− 1
−s − sin at e−s t dt
s 0
a 1
− as f¯(s)
= s s
(293)
Rearranging, we obtain s2 f¯ = a − a2 f¯ ⇒ f¯ = a/ s2 + a2 as required.
5.1 Properties
The following properties of the Laplace transform are easily shown from its definition.
52
1. Linearity: If a and b are constants, then
Z ∞ Z ∞ Z ∞
(af1 (t) + bf2 (t)) e−st dt = a f1 (t)e−st dt + b f2 (t)e−st dt (294)
0 0 0
so that:
L [af1 + bf2 ] = af¯1 + bf¯2 (295)
2. Powers of t:
n!
L [tn ] =
. (296)
sn+1
This follows form L [1] = 1/s using proof by induction:
Z ∞
n
L [t ] = tn e−st dt
0
−st ∞ Z ∞
e e−st
= tn − ntn−1 dt
−s 0 0 −s
Z ∞
n
= tn−1 e−st dt
s 0 (297)
n n−1
= L t
s
nn−1n−2 21 0
= ... L[t ]
s s s ss
nn−1n−2 21 1 n!
= ... = n+1 .
s s s ss s s
3. Shifting in s:
L eat f (t) = f¯(s − a)
(298)
Example: L[eat ].
If f (t) = 1 then f¯(s) = 1
s from (292) so L[eat ] = L[eat .1] = f¯(s − a) = 1
s−a .
dn ¯
L [tn f (t)] = (−1)n f (s) (299)
dsn
5. Shifting in t:
L [f (t − t0 )θ(t − t0 )] = e−t0 s f¯(s) (300)
where θ is the Heaviside “step function”, defined by
0, t ≤ 0,
θ(t) = (301)
1, t > 0,
which ensures that we only consider f (t) for non-negative values of t, even when there is a
time delay t0 .
6. Derivatives in t become powers of s:
n
d
L f (t) = sn f¯(s) − sn−1 f (0) − sn−2 f 0 (0) − ... − f (n−1) (0) (302)
dt
Here f (k) denotes the k th derivative of f (t) at t = 0. The Laplace transform is therefore
particularly suitable for initial value problems: problems in which initial conditions on the
function and its derivatives are given at t = 0.
53
7. Integration in t: Z t
1¯
L f (τ ) dτ = f (s) (303)
0 s
8. Convolution: Z t
L [(f ∗ g)(t)] = L f (t − τ )g(τ ) dτ = f¯(s)ḡ(s) (304)
0
5.2 Proofs
To prove the convolution theorem (8), we first form the product of the defining integrals for f¯(s)
and ḡ(s): Z ∞Z ∞
0
f¯(s)ḡ(s) = e−s(τ +τ ) f (τ 0 )g(τ )dτ 0 dτ. (306)
0 0
Next we introduce a change of variables, letting τ 0 = t − τ and use the fact that f vanishes for
negative values of its arguments. Thus
Z ∞ Z t
f¯(s)ḡ(s) = e−st f (t − τ )g(τ )dτ dt. (307)
0 0
To prove the periodicity property (9), note that for a function f (t) such that f (t + T ) = f (t)
Z ∞ ∞ Z
X (n+1)T
f¯(s) = e−st f (t)dt = e−st f (t)dt (310)
0 n=0 nT
and hence
∞
X Z T Z T
f¯(s) = e−nsT e−sτ f (τ )dτ = (1 − e−sT )−1 e−st f (t)dt, (311)
n=0 0 0
where the first equality follows from the change of variable t = nT + τ and the second equality
follows from the known result for the sum of a geometric series and the change of variable t = τ .
The inverse formula using the Bromwich contour can be proved using the inverse R∞Fourier
transform. The main ingredients of the proof are letting s = a + iy, and replacing 0 dt by
R∞
−∞
H(t) dt.
5.3 Uses
Laplace transforms allow us to convert complicated integro-differential equations into simpler prob-
lems. √
Example: Consider the following equation, with boundary conditions f (0) = 0 and f (π) = 2:
Z t
df (t)
f (τ )dτ + 4tf (t) = . (312)
0 dt
54
Taking the Laplace transform we get
1 ¯
s f (s) + −4 dds f¯(s) = sf¯(s) − f (0)
(by 7) (by 1 and 4) (by 6) (313)
This is a 1st-order ODE for f¯(s). Using the first initial condition it becomes
d ¯
f (s) = f¯(s) 1s − s
4 ds
4 ¯ 1
f¯
df = s − s ds
(314)
2
4 ln f¯ + C = ln(s) − s2
2
f¯4 = c s e−s /2
1 2
f¯ = c̃ s 4 e−s /8
we have one constant of integration, c, to be determined by using the remaining boundary condition.
This method is only useful if we can get back to f (t) once we have found our Laplace transform
f¯(s). There are two ways of doing this, by:
inspection or
inversion (theorem)
Example: Find f by inspection, given f¯ = 4 2(s − 7)3 + 9s2 + 36 / (s2 + 4)(s − 7)3 .
The inversion theorem guarantees that any continuous and piecewise smooth function is uniquely
determined by its transform.
We can choose any real number a such that the line γ from a − i∞ to a + i∞ lies to the right
of all the singularities of f¯(s). Since f¯(s) is analytic in Re s > α, a > α will do.
This is called the Bromwich integral and, if f (t) is continuous, it takes the form
Z a+iR
1
f (t) = lim f¯(s)est ds, t > 0. (317)
2πi R→∞ a−iR
If f¯(s) has no branch points, we can evaluate the Bromwich integral by closing the contour
with a large semicircle to the left. This is called the Bromwich contour. But for this we need
to be able to show that the contribution of the large semicircle vanishes as its radius R → ∞.
55
Im γ a + iR
R
R
γ
Γ
α a Re
a − iR
If there is a branch cut going to −∞ the contour can be wrapped around the branch cut, and
the contribution from the integral on both sides of the branch cut needs to be taken into account
explicitly.
We can now apply the Residue Theorem to the Bromwich contour:
Z X
f¯(s) est ds = 2πi Res(f¯(s) est : ak ). (318)
Γ
singularities ak
of f¯(s)est
inside Γ
Note that the ak ’s that appear in the formula are the residues of est f¯(s), not f¯(s). In the limit
R → ∞ all the singularities will be inside Γ. We can also write the integral as
Z Z a+iR Z
f¯(s) est ds = lim f¯(s) est ds + lim f¯(s) est ds. (319)
Γ R→∞ a−iR R→∞ γR
Hence: Z a+iR
2πi f (t) = lim f¯(s) est ds
R→∞ a−iR
Z Z
= f¯(s) est ds − lim f¯(s) est ds (320)
Γ R→∞ γR
X Z
= 2πi Res(f¯(s) est : ak ) − lim f¯(s) est ds.
R→∞ γR
ak
If we can show that the contribution from the large semicircle γR vanishes as its radius R goes to
∞, then we have f (t) in terms of residues which we can calculate.
The following Lemma tells us when we can neglect the contribution from the large semicircle:
Lemma (Bromwich contour lemma) If there exist constants K > 0, k > 0 and R0 > 0
such that
K
|f¯(s)| < k (321)
R
56
for all s on the large semicircle to the left |s| = R, Res < a, for all R > R0 , then
Z
f¯(s) est ds → 0, as R → ∞. (322)
γR
Im
a + iR
γR
Re
−i
a − iR
Figure 38: The Bromwich contour for f¯(s) = (2 − s)/(s2 + 1) requires a > 0.
π 3π
On γR , s = a + Reiθ , where 2 ≤θ< 2 , so
2 − a − Reiθ R+2+a K
|f¯(s)| = 2 iθ 2 2iθ
≤ 2 2
< (325)
1 + a + 2aRe + R e R − 2aR − a − 1 R
for some K > 1 and R > R0 for some R0 > 0. Note that we need to fix R0 as least big enough
so that the numerator and denominator in the second fraction here are both positive. K then
depends on how much bigger than that we make R0 . The sloppy reasoning would be that
1
|f¯(s)| → as R → ∞, (326)
R
But for the Lemma we need a rigorous inequality, and for this we need to fix a suitable R0 , and it
is clear that we will need K > 1 for any R0 .
57
Hence
X 2 − s st
f (t) = Res e : ak
s2 + 1
singularities ak
of f¯(s)est
2−s 2−s
= Res (s+i)(s−i) est : i + Res (s+i)(s−i) est : −i
+it
+e−it 2 e+it −e−it
= −e 2 + i 2
= − cos t + 2 sin t.
Example: Find f (t) by inspection, given f¯(s) = (2 − s)/(s2 + 1).
2−s 1 s
f (t) = L−1 = 2L −1
− L −1
= 2 sin t − cos t. (328)
s2 + 1 s2 + 1 s2 + 1
1
Example: Find f (t) given f¯(s) = s− 2 .
1
f¯(s) = √ has a branch point at s = 0 but no other singularities. (329)
s
Im γR + a + iR
γ+ γε
γ− Re
Γ
γR−
a − iR
58
Since Z Z Z Z Z Z Z
= + + + + +
+ −
Γ γ γR γ+ γ γ− γR
Z Z Z Z Z (332)
0 = 2πif (t) + + + + +
+ −
γR γ+ γ γ− γR
Calculating both integrals on either side of the branch cut together (this may prove useful, as there
often exists symmetry) we get
Z Z
f¯(s) ds + f¯(s) ds
γ+ γ−
0 ∞
e−xt e−xt
Z Z
= √ (−dx) + √ (−dx)
∞ i x 0 −i x
(336)
∞ ∞
e−xt e−xt
Z Z
= −i √ dx + −i √ dx
0 x 0 x
∞
e−xt
Z
= −2i √ dx
0 x
√
If we let X 2 = x then X = x and dx = 2X dX. To substitute we need to change limits:
x = 0 ⇒ X = 0, x = ∞ ⇒ X = ∞.
∞ ∞ 2 ∞
e−xt e−X t
Z Z Z
2
−2i √ dx = −2i 2X dX = −4i e−X t dX
0 x 0 X 0
Z ∞
2 (337)
= −2i e−X t dX (using symmetry)
∞
pπ
= −2i t (standard result)
59
Finally we have, remembering (333):
r
−1 π 1
f (t) = −2i + 0 = √ . (338)
2πi t πt
Differentiating with respect to x may also be taken into or out of the integral, so that
The properties of Laplace transforms (#6) tell us that the transforms of time derivatives are
given by
L [ut (x, t)] = sū(x, s) − u(x, 0),
(342)
L [utt (x, t)] = s2 ū(x, s) − su(x, 0) − ut (x, 0).
These results can of course be extended to higher order, and used to find mixed derivatives such
as
L [uxt (x, t)] = sūx (x, s) − ux (x, 0). (343)
This way, a second-order partial differential equation can be transformed into a second-order
ordinary differential equation (ODE). Very often it is easier to solve the ODE and then transform
back from ū(x, s) to u(x, t) rather than solve the PDE directly. Notice that the Laplace transform
of ut (x, t) brings in u(x, 0), the Laplace transform of utt (x, t) brings in u(x, 0) and ut (x, 0), and so
on. That is why Laplace transform methods are so suitable for dealing with initial value problems.
To solve a PDE using Laplace Transforms we need to:
I Transform: both the equation and boundary conditions;
II Solve: the resulting ODE for ū(x, s);
III Invert: To find u(x, t), the solution to the original problem.
Example: Solve ux = 2ut + u for t, x > 0, given u(x, 0) = e−3x and u(x, t) → 0, x → ∞.
I Transform equation:
ux = 2ut + u
ūx = 2 (sū − u(x, 0)) + ū (344)
ūx − (1 + 2s)ū = −2e−3x
60
II Solve ODE in x:
d h −(1+2s)x i
ū e = −2e−3x e−(1+2s)x (apply integration factor)
dx
Z (To treat this as an ODE in x
ū e−(1+2s)x + C(s) = −2e−(2s+4)x dx requires constants of integration
to be functions of s) (346)
2e−(2s+4)x
ū e−(1+2s)x = − − C(s)
−(2s + 4)
−3x
e
ū(x, s) = − C(s) e+(1+2s)x
s+2
We cannot invert the solution yet, because C(s) is an unknown function of s which will affect the
dependence of u on t. To obtain C(s) we next use the BC:
II Solution of ODE with BC:
−3x
e
0 = lim ū(x, s) = lim − C(s) e+(1+2s)x = 0 + C(s) lim e(2s+1)x . (347)
x→∞ x→∞ s + 2 x→∞
e(2s+1)x → 0 as x → ∞ if and only if Re(2s) > −1, which does not hold for all s. Hence, for
boundary condition to be satisfied, we must have C(s) = 0.
III Invert: −3x
ū(x, s) = es+2 h i
u(x, t) = e−3x L−1 s+2 1
(348)
u(x, t) = e−3x e−2t
u(x, t) = e−3x−2t
Example: Solve utt = uxx , for 0 < x < l, t > 0 with Note:
initial conditions : u(x, 0) = 0, ut (x, 0) = sin πx
l 2 I.C.s as 2nd order in t
and boundary conditions: u(0, t) = 0, u(l, t) = 0. 2 B.C.s as 2nd order in x
ū = α sin πx πx
l + β cos l d2 ū
π πx π − s2 ū =
ūx = l α cos l − l β sin πx
l ⇒ 2 2
dx2
π2 π2 − πl2 α sin πx
l −
π
β cos πx
− s2 α sin πx 2 πx
l − s β cos l
ūxx = − l2 α sin l − l2 β cos πx
πx
l
l2 l
(350)
Hence, π 2 πx πx πx
−s2 − α sin + β cos = − sin (351)
l l l l
π 2
s2 + l α = 1 α = 1
2
⇒ s2 +( πl ) (352)
π
2
s2 + l β = 0 β = 0
sin πx
ū(x, s) = A(s)esx + B(s)e−sx + l
π2
. (353)
s2 + l2
Now impose boundary conditions:
ū(0, s) = 0 = A(s) + B(s), ū(l, s) = 0 = A(s)els + B(s)e−ls ,
61
⇒ A(s) = −B(s) ⇒ A(s) els − e−ls = 0 ⇒ A(s) = 0 ⇒ B(s) = 0.
III Invert:
By inspection: " #
sin πx
u(x, t) = L−1 l
2
s2 + π"l2 #
πx −1 1
u(x, t) = sin L 2
l s2" + πl2 # (355)
π
l πx −1 l
u(x, t) = sin L 2
π l s2 + πl2
l πx πt
u(x, t) = sin sin .
π l l
est est
−1 1 π π
L 2 = Res : + i + Res :− i
s2 + πl2 (s − iπ/l)(s + iπ/l)
i(π/l)t −i(π/l)t
i π t l −i π t (s − iπ/l)(s + iπ/l) l
e e l e l −e l l πt
= + = = sin .
2iπ/l −2iπ/l π 2i π l
(356)
Example: Solve ut = uxx , for 0 < x < 1, t ≥ 0 with
boundary conditions: ux (0, t) = 0, u(1, t) = u1 (a constant), 2 B.C.s as 2nd order in x
and initial condition: u(x, 0) = u0 (a constant). 1 I.C. as 1st order in t
u1 /s − u0 /s u1 − u0
⇒ A(s) = B(s) = √ √ = √ √ . (358)
e s + e− s s e s + e− s
Hence,
√ " √ #
ex s + e−x s
u1 − u0 u0
ū = √ √ + (359)
s e s + e− s s
√
u1 − u0 cosh sx u0
= √ + . (360)
s cosh s s
h √ i
III Invert: u(x, t) = u0 L−1 1s + (u1 − u0 ) L−1 scosh sx
√
cosh s
.
62
By inspection: L−1 1s = 1.
√
Consider the inversion of f¯(x, s) = cosh √ sx
s cosh s
using the Complex Inversion Theorem. First check
that the theorem is applicable:
√ √
1 ex√ R
f¯(x, s) = cosh √ sx
∼ R e R ≤ M R−K
s cosh s (361)
(On arc, for R 1) (for K = 1 and since x< 1)
Hence X
f (x, t) = Res(f¯(x, s) est : ak ). (362)
singularities ak
of f¯(x, s)est
√
The singularities
√ of√f¯(s) are a simple pole at s = 0 and the roots of cosh s.
2n+1 z
cosh s = 0 ⇒ s = 2 πi, n ∈ Z, (Remember cosh z = cos i )
2n+1 2 2
+
⇒ s = an = − 2 π ,n∈Z . √ √
We take only n ≥ 0 as the values of an for n < 0 would be repetitive. Note cosh z, unlike z, is
single valued:
√ 2 √ 4
√ ( s) ( s) s s2
cosh z = 1 + + + ... = 1 + + + ... (363)
2! 4! 2! 4!
Note that within our√ boundaries, 0 < x < 1, the above singularities do not cancel any root of
the numerator, cosh x s (remember that we are interested in singularities with respect to s; not
x or t).
The singularities an are all simple poles, with residues given by
√
est cosh x s
Res(f¯est : an ) = lim √s √ (using hg0 formula) (364)
s→an sinh s
2
Hence: 2
q
2n+1
π2 t
e ( )
− 2n+1 2
2n+1 2
2
cosh x − 2 π2
Res(f¯est : −
2 π2 ) = 2q q
2n+1 2 2 2n+1 2
− 2 π sinh − 2 π2
cosh 2n+1
−( 2n+1
2
) π2 t 2 πxi
= 2e 2 2n+1 2n+1
2 πi sinh 2 πi (365)
cos 2n+1
2n+1 2 2
2 πx
= 4e−( 2 ) π t
− (2n + 1)π sin 2n+1
2 π
2n+1
4 2 2 cos πx
−( 2n+1 ) π t 2
= − e 2
n
(2n + 1)π (−1)
√
cosh x s st
For the simple pole s = 0 we have: Res(f¯est : 0) = lim √ e = 1.
s→0 cosh s
Substituting these values into the Complex Inversion Theorem formula (362) gives:
∞
(−1)n+1 4
2n + 1 2n+1 2 2
πx e−( 2 ) π t .
X
f (x, t) = 1 + cos (366)
n=0
(2n + 1)π 2
63
It is useful to check that our solution satisfies all the necessary conditions:
2
(−1)n+1 2n+1
π2 t
sin (0) e−( )
2n+1
P∞
ux (0, t) = 0 − 2 π 4(u1 − u0 ) n=0 (2n+1)π
2 = 0;
(369)
2
(−1)n+1 2n+1
π2 t
e−( )
P∞ 2n+1
u(1, t) = u1 + 4(u1 − u0 ) n=0 (2n+1)π cos 2 π
2 = u1 .
for all values of x ∈ (0, 1). That this indeed holds can be confirmed numerically—see Figure 40.
1
0.785
0.5
−0.5
−0.785
−1
−1 0 1 2 3
P10 P100 x
Figure 40: The sum (372) for 0 (dotted) and 0 (solid). In the relevant domain (0 < x < 1)
the sum seems to converge to ∼ 0.785 ∼ π/4.
The expression (372) is in fact a Fourier series solution of a square wave with amplitude π4 ,
extended as an even function. This result suggests that our solution (368) is a Fourier Series of
a much simpler (closed form) expression for u(x, t). The inversion of Fourier series solutions is
beyond this course, but it is worth noting that Transform methods and Series solutions are closely
related (see section 4.2).
5.6 Applications
In this section we briefly discuss physical problems that can be described by partial differential
equations of the types analysed in the previous section.
Heat conduction: Consider a solid located at x ≥ 0, i.e. with a boundary at x = 0. Let
the solid be initially (time t = 0) at a temperature Ti and suppose that the boundary x = 0 is
maintained at a temperature T0 for time t > 0. The rate at which heat is transferred across a
64
plane section, of constant x, in the direction of increasing x is −KTx , where K is the thermal
conductivity and Tx is temperature gradient.
The rate at which heat is transferred into a slab bounded by x and x + dx is KTxx per unit area
(in the yz plane). This rate must be equal to the rate at which the slab is gaining heat, namely
ρCTt dx, where ρ is the density, C is the specific heat and dx is the volume per unit of transverse
area. Equating these two rates we obtain the diffusion equation
Tt = κTxx (373)
where
K
κ= (374)
ρC
is called the diffusivity.
This equation is indeed of the form analysed in the previous section. We are given initial
conditions for all x > 0, i.e. T = Ti at t = 0, and we are given a boundary condition T = T0 at
x = 0. Since the equation has two spatial derivatives, we will need a second boundary condition in
x. We can consider either a finite size slab, in which case we will often be told the temperature at
the second boundary x = L, or an infinite slab, in which case we will usually be told that T → 0
as x → ∞.
Wave propagation along bars: Consider a uniform bar of cross section A, extended along
the x axis. The bar has one end fixed at x = 0 and has length l. Suppose a stress is applied to the
bar in the direction of the positive x axis, causing the bar to make small longitudinal displacements
h(x, t) along the x axis. These displacements satisfy a wave equation
where
E
c2 = (376)
ρ
with E the Young modulus and ρ the density. Again this equation is of a type analysed in the
previous section. To solve the equation we will need two initial conditions (h(x, 0) and ht (x, 0))
and two boundary conditions. One boundary condition follows from the fact the end is fixed at
x = 0, i.e. h(0, t) = 0; the second boundary condition might be ht (0, t) or ht (l, t).
Wave propagation along strings: Consider a string extended along the x axis. Suppose
that a stress is applied, causing the string to make small displacements in a transverse direction,
i.e. y(x, t) . Again these displacements satisfy a wave equation
6 Hankel transforms
The complex Fourier transform and Laplace transform are by far the two most used integral
transforms, but there are many other integral transforms that are useful for transforming PDEs
into ODEs. Here we give only one more example: For problems in cylindrical polar coordinates we
can often use Hankel transforms. Their integral kernel is not simply e−st or e−ikx , but the more
complicated Bessel functions.
65
6.1 The Bessel equation
Consider the wave equation in two space dimensions:
When we go from the Cartesian coordinates x and y to cylindrical polar coordinates r and ϕ,
defined by
x = r cos ϕ, y = r sin ϕ (380)
then the wave equation transforms into
1 1
ftt = frr + fr + 2 fϕϕ . (381)
r r
Now we can use the technique of separation of variables to write f (r, ϕ, t) as a product of three
functions of one variable each:
f (r, ϕ, t) = eiωt einϕ u(r) (382)
where ω is a real constant and n is an integer. We find that u(r) obeys the ODE
d2 u 1 du n2
2
+ + ω − 2 u=0 (383)
dr2 r dr r
This seems to depend on two parameters ω and n but if we make the substitution ωr = x we
obtain the simpler equation
d2 u 1 du n2
+ + 1− 2 u=0 (384)
dx2 x dx x
This equation is called the Bessel equation of order n.
The constants Cn (ω) are arbitary. Note that if the problem is cylindrically symmetric, so that f
depends only on r and t, then we only need the term with n = 0 in the series, so that
Z ∞
f (r, t) = C0 (ω)eiωt J0 (ωr) dω. (387)
−∞
66
6.3 The Hankel transform
The Hankel transform of order n of a function f (r) is defined by
Z ∞
˜
Hn f (r) = fn (α) = Jn (αr)f (r)rdr. (388)
0
The kernel of this transform is therefore rJn (αr).
One can show that the Hankel transform is its own inverse:
Z ∞
f (r) = Jn (αr)f˜n (α)α dα, (389)
0
for any integer n.
If we now transform to polar coordinates, then we expect this to break up into a Fourier series and
a Hankel transform. So define
x1 := r cos θ, x2 := r sin θ, (391)
k1 := α cos φ, k2 := α sin φ. (392)
In a slight abuse of notation, define fˆ(k1 , k2 ) = fˆ(α, ϕ), and f (x1 , x2 ) = f (r, θ). Define the Fourier
series coefficients fn (r) by
X∞
f (r, θ) =: fn (r)einθ . (393)
−∞
Then one can show that
∞
X
fˆ(α, ϕ) = fˆn (α)einφ , (394)
−∞
where
fˆn (α) = (−i)n Hn [fn (r)] (α). (395)
67
Example Solve
1
ftt (r, t) = frr (r, t) + fr (r, t) (398)
r
for f (r, t) for r ≥ 0 and t ≥ 0 subject to the following initial conditions:
Implicitly we use the boundary conditions when applying this expression. Transforming the left
hand side of the equation is straightforward, as t derivatives commute with the transform:
∂2 ˜
H0 (ftt (r, t)) = f 0 (α, t) . (402)
∂t2
Thus the transformed equation is an ordinary differential equation in time:
∂2 ˜
f 0 (α, t) = −α2 f˜0 (α, t) (403)
∂t2
We also need to transform the initial conditions.
while
∂ ˜
f0 (α, 0) = H0 e−kr .
(405)
∂t
The latter can be shown to be
k
3 . (406)
(k 2 + α2 ) 2
We can now solve the transformed equation subject to these initial conditions. The general solution
of the transformed equation is
68
7 Nyquist stability theory
7.1 Stability of circuits
In this section we will apply the theory of Laplace transforms to circuits.
Suppose an input voltage Vin (t) and an output voltage Vout (t) of a circuit are related by the
ordinary differential equation
(n) (n−1) (1) (m)
an Vout (t) + an−1 Vout (t) + ... + a1 Vout + a0 Vout (t) = bm Vin (t) + ... + b0 Vin (t), (410)
where n > m and an , bm are constants. Suppose that Vin and Vout , and all their time derivatives
up to (not including) order n vanish at t = 0. Taking the Laplace transform of equation (410) and
using the derivative property of the Laplace transform, we find that
an sn V̄out (s) + an−1 sn−1 V̄out (s) + ... + a1 s V̄out (s) + a0 V̄out (s) = bm sm V̄in (s) + ... + b0 V̄in (s).
(411)
We can rearrange this to get
Ā(s) is called the system transfer function. From the convolution property of the Laplace
transform Z t
−1
V̄out = V̄in Ā(s) ⇒ Vout (t) = L [V̄in Ā] = A(t − τ )Vin (τ ) dτ. (413)
0
We see that A(t) is the reaction of the circuit to a δ-function blip in the input at t = 0. Because of
linearity, the reaction of the circuit to an arbitrary input Vin (t) is just an integral over such blips.
The function A(t − τ ) is known as the Green function of the circuit.
This method is useful for determining the reaction to an arbitrary input, but here we shall
only ask if the circuit is stable or not. “Stable” here means that a bounded input gives a bounded
output, and this, in turn, means that A(t) does not blow up as t → ∞.
We shall now determine the stability of the circuit based on Ā(s).
Because Ā(s) is a rational function, the residue of est Ā(s) at s = a is eat times a polynomial
in t. (The order of the polynomial is the order of the pole minus 1, as can be shown by the
derivative rule formula for the calculation of residues.) Therefore, poles in Ā(s) with Re s < 0
give contributions to A(t) that decay exponentially at late times, while poles with Re s > 0 give
contributions that blow up exponentially at late times and make the circuit unstable.
Poles on the imaginary axis are critically stable and require further investigation. Simple poles
on the imaginary s-axis give contributions to A(t) that are constant or oscillate at late times -
these are marginally stable. Higher poles on the imaginary axis give blowup as a power of t and
hence are unstable.
Henceforth we assume that all critically stable poles are marginally stable, since the case when
there exists an unstable pole on the imaginary axis is trivial (the circuit is unstable!):
The circuit is stable if Ā(s) has no poles in the right hand complex plane.
69
Im
stable 1111111111111
0000000000000
0000000000000
1111111111111
0000000000000
1111111111111
0000000000000
1111111111111
0000000000000
1111111111111
unstable
0000000000000
1111111111111
0000000000000
1111111111111
0000000000000
1111111111111
0000000000000
1111111111111
0000000000000
1111111111111
0000000000000
1111111111111
0000000000000
1111111111111Re
0000000000000
1111111111111
0000000000000
1111111111111
0000000000000
1111111111111
0000000000000
1111111111111
0000000000000
1111111111111
0000000000000
1111111111111
0000000000000
1111111111111
Figure 41: Stabilizing and destabilizing poles.
Vfb Vout
where we used V̄out (s)/V̄in (s) = Ā(s). Noticing that V̄fb (s)/V̄in (s) = (V̄fb /V̄out )(V̄out /V̄in ) =
Ā(s)B̄(s), we obtain the total transfer function of the system as
Ā(s)
f¯(s) = . (416)
1 + Ā(s)B̄(s)
The loop is stable if f¯(s) has no poles in the right hand plane.
It is usually assumed that both systems A and B are stable on their own—only then does it
make sense to ask about stability of the integrated feedback loop. We henceforth thus assume that
neither Ā(s) nor B̄(s) have poles in the right hand plane. In this case, the only poles of f¯(s) are
possible zeros of
ḡ(s) ≡ 1 + Ā(s)B̄(s). (417)
[Note that zeros of ḡ(s) cannot cancel out with zeros of the numerator, Ā(s), if B̄(s) has no poles.]
The feedback circuit will be stable if and only if ḡ(s) has no zeros in the right hand plane.
If ḡ(s) is a simple enough polynomial, we can find its zeros directly. In more complicated cases
we can use the Principle of the Argument, with a suitable contour, to determine the number of
zeros ḡ(s) has in the right half plane. (For our purpose we don’t actually need to find those zeros,
only to tell if there exist any!)
70
is then closed at infinity with a semicircular arc (γR ), such that the contour covers the entire right
hand plane as the radius goes to infinity (see figure 43). The closed contour is usually denoted C.
γR
All the zeros and poles in the right hand plane will be inside the contour C. If there is a zero
or pole on the imaginary axis we exclude it with an -semicircle (see Figure 44): we wish to only
count unstable poles as critically stable poles are considered separately.
Im
ε −arc
Re
The principle of the argument states that, for a contour C and function ḡ(s)
W = N − P, (418)
where N = number of zeros of ḡ inside C,
P = number of poles of ḡ inside C, and
W = number of times the ḡ(s) winds around the origin in the ḡ (ḡs = 0) as s moves around the
contour C.
Recall that for the feedback circuit considered above we have stability if and only if ḡ(s) =
1 + Ā(s)B̄(s) has no zeros in the right hand plane. Namely, the circuit is stable if and only if
N = P + W = 0. (419)
71
If both systems A and B are stable, Ā(s) and B̄(s) have no poles in the right hand plane. In
this case, ḡ(s), too, has no poles in the right hand plane, and we have P = 0. (In solving problems
it is necessary to first check that A and B are stable before assuming P = 0!)
If P = 0 then the stability criterion becomes simply
W =0 (420)
In this case, one therefore only needs to calculate the winding number of the contour C.
Most textbooks define
h̄(s) = ḡ(s) − 1 = Ā(s) B̄(s), (421)
and then ask how many times h̄(s) winds around h̄(s) = −1 as s goes around C, instead of asking
how many times ḡ(s) winds around ḡ(s) = 0 as s passes around the contour C.
To answer this question one should sketch the path h̄(C), as in the following example.
Example: Determine the stability of a feedback system formed from a circuit with transfer func-
tion Ā(s) = 1−s k
1+s and a feedback circuit with transfer function B̄(s) = s where k > 0 is a (real)
constant.
γ+
γ
R
2 γ
ε
3
γ−
k(1−s)
Figure 45: The Nyquist contour used for h̄(s) = s(1+s) .
(1 − Reiθ )k ke−iθ
h̄(γR ) = lim = lim − = 0. (423)
R→∞ (1 + Reiθ )Reiθ R→∞ R
(So the image of h̄(s) along the entire semi-circle γR shrinks to a point.)
72
On γ+ we have s = iω with ω > 0. Then,
(1 − iω) 2k k(ω 2 − 1)
h̄(γ+ ) = k =− 2
+i . (424)
iω(1 + iω) 1+ω ω(ω 2 + 1)
At point 1 (ω → +∞) we have already found h̄ = 0.
At point 2 (ω → 0) we have πh̄ →π−2k − ωk i., which is an asymptote x = −2k with y → −∞i.
−iθ
On γ , s = e for θ ∈ − 2 , + 2 , since we move clockwise around the semicircle, so:
k 1 − e−iθ k
h̄(s) = −iθ −iθ
∼ e+iθ for 1. (425)
e 1 + e
Hence h̄(γ ) is still a semicircle, but orientated anticlockwise from − π2 to + π2 and with radius
k
→ ∞ as → 0.
Finally, on γ− we have s = −iω with ω > 0. We notice then from Eq. (424) that h̄(γ− ) is
complex conjugate to h̄(γ+ ). This means that the image of h̄(γ− ) can be obtained from the image
of h̄(γ+ ) by merely reflecting through the x axis.
To be able to plot the image of h̄(s) it is also useful to look at points where h̄(s) crosses the
axes:
h̄(γ± ) crosses the real axis when
k(ω 2 − 1)
Im h̄(s) = ± = 0, (426)
ω(ω 2 + 1)
2k
i.e., when ω = 1, for which x = − 1+1 2 = −k.
Hence h̄ cuts the real axis at h̄(s) = −k when s = ±i.
h̄(γ± ) crosses the imaginary axis when
2k
Re h̄(s) = − = 0, (427)
1 + ω2
i.e., when ω = ±∞, for which y = 0.
Hence h̄ cuts the imaginary axis at h̄(s) = 0 when s = ±i∞.
We are now ready to sketch h̄ (C)—see
figure 46.
2, if −k < −1,
Our sketch indicates that W = Hence
0, if −k > −1.
2, if k > 1,
N= (428)
0, if k < 1.
The circuit is therefore stable if and only if k < 1. Combining these results we see that the circuit
is only stable for 0 < k < 1.
Example: Determine the stability of the above closed loop circuit by considering the poles of f¯
directly.
1−s
Ā(s) 1+s s(1 − s)
f¯(s) = = k 1−s
= 2 . (429)
1 + Ā(s)B̄(s) 1 + s 1+s s + (1 − k)s + k
The poles of f¯ are the zeros of
s2 + (1 − k)s + k. (430)
The roots are given by
1 p
s± = k − 1 ± 1 − 6k + k 2 , (431)
2
and the circuit is stable provided the real parts of s+ and s− are both not positive.
Fig. 47 shows a plot of Re(s± ) as a function of k. The plot was obtain by evaluating Eq. (431)
numerically. The plot confirms our finding that the system is only stable when 0 < k < 1.
In examples where the transfer function f¯ is more complicated, it may be very difficult to
obtain the poles of f¯ directly, as in the last example. In such cases, applying the Nyquist stability
analysis may require much less work and will usually be more effective.
73
h(Γ)
asymptote
1
A k/ ε
−2k −k h(i )
h(γ ε)
h(γ+)
−1
2
−i B −i
Figure 46: The path of h̄(s) in the complex plane as s travels along C.
Re(s)
s−=s+=1−21/2 distinct
real roots
2 when k=3−23/2
Re(s+) Re(s−)
0
distinct real roots
Re(s )=Re(s )
+ −
−2 Complex conjugate roots
Re(s )
−
−4
−3 −2 −1 0 1 2 3 4 5 6 7 8
k
74