IFT LectureNotes
IFT LectureNotes
Notes taken by
Richie Dadhley
richie@dadhley.com
i
Acknowledgements
These are my notes on the 2019 lecture course "Introduction to Field Theory" taught by
Professor Michael Spannowsky at Durham University as part of the Particles, Strings and
Cosmology Msc. For reference, the course lasted 4 weeks and was lectured over 24 hours.
I have tried to correct any typos and/or mistakes I think I have noticed over the course.
I have also tried to include additional information that I think supports the taught material
well, which sometimes has resulted in modifying the order the material was taught. Obviously,
any mistakes made because of either of these points are entirely mine and should not reflect
on the taught material in any way.
I have also used/borrowed from Professor David Tong’s notes throughout the course (par-
ticularly at the start). This notes are a must for any new students to QFT and can be found
via
http://www.damtp.cam.ac.uk/user/tong/qft/qft.pdf
The material is presented in the order taught. This generally coincides with Professor
Tong’s notes, however there are small deviations here and there. Probably the biggest dif-
ference is the fact that we present the LSZ theorem in this course, whereas that is saved for
the Advanced QFT course at Cambridge. Equally, normal ordering is not introduced until
lecture 7, where Wick’s theorem is discussed, whereas Prof. Tong introduces it much earlier
(in order to drop the infinities in expectation values).
I would like to extend a message of thanks to Professor Spannowsky for teaching this
course. I would also like to thank Professor Tong for uploading his notes to the internet.
If you have any comments and/or questions please feel free to contact me via the email
provided on the title page.
https://richie291.wixsite.com/theoreticalphysics
R˚i`c‚h˚i`e D`a`d˛hffl˜l´e›y
Contents
ii
CONTENTS iii
7 Wick’s Theorem 53
7.1 Normal Ordering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 53
7.2 Wick’s Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
7.2.1 Recovering The Propagator . . . . . . . . . . . . . . . . . . . . . . . . 54
7.2.2 Wick Contractions & Wick’s Theorem . . . . . . . . . . . . . . . . . . 55
7.2.3 Wick’s Theorem For States . . . . . . . . . . . . . . . . . . . . . . . . 57
7.3 Scalar Yakawa Scattering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
7.3.1 Fully Connected . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 58
7.3.2 Feynman Rules — Position Space . . . . . . . . . . . . . . . . . . . . . 59
7.3.3 Feynman Rules — Momentum Space . . . . . . . . . . . . . . . . . . . 61
10 Dirac Fields 82
10.1 Representations Of Lorentz Group . . . . . . . . . . . . . . . . . . . . . . . . 82
10.1.1 The Generators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83
10.1.2 The Dirac Representation . . . . . . . . . . . . . . . . . . . . . . . . . 84
10.1.3 Spinors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86
10.1.4 Non-Unitarity Of S[Λ] . . . . . . . . . . . . . . . . . . . . . . . . . . . 87
10.2 The Dirac Action & Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . 87
10.2.1 Obtaining The Klein-Gordan Equation . . . . . . . . . . . . . . . . . . 90
10.2.2 γ 5 & Dirac Bilinears . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90
10.2.3 Chiral Spinors & Projections . . . . . . . . . . . . . . . . . . . . . . . 92
The first question we should ask ourselves is "why does quantum field theory (QFT) even
exist and what is it used for?" The answer to this question comes from the fact that special
relativity and quantum mechanics (QM) don’t get along, and we need some method to make
them consistent. Let’s see why.
1.1.1 Causality
In the pre-Einstein ages, we often just had formulae for the forces between objects and just
got on with using them. For example, let’s say we have two particles of charges q1 and
q2 . Let’s plonk them down somewhere and label their separation by d. Then the rules of
electromagnetism tell us that the force between these two particles is given by
q1 q2
FEM ∝ 2 . (1.1)
d
We were happy with this, until Einstein came along and told us about a little thing called
causality. This is basically the statement that information cannot travel faster than the speed
of light. In the language of special relativity that is ‘information cannot travel between two
spacelike separated points’. This wreaks havoc for Equation (1.1). Why? Well because there
is no time dependence! Therefore if we were to move one of the charges, q1 say, then the
value of d would change, and Equation (1.1) tells us that q2 would feel this change in force
instantly. In other words, the information contained in the fact that we have moved q1 is
transmitted to q2 (which could be arbitrarily far away) instantaneously; not good! We can
fix this problem if instead we think of the force being mediated by some field that travels at
the speed of light, and this is exactly what we do.
Remark 1.1.1 . Note that when q1 and q2 are close by, the fact that light travels so fast
means that the difference between the result obtained by Equation (1.1) and the field theory
description will be negligible. This is the usual story of non-relativistic vs. relativistic theories.
Ok, that’s all very well and good, but what on earth does that have to do with QM? Well
in quantum mechanics we have operators which act on the states of our Hilbert space and do
something to them. This is information! If we couple this idea with the fact that we want to
be able to act with operators at any point in space and time1 we see that we are in a very
similar situation to the electromagnetism case.
1
From now on I shall only say spacetime, and not ‘space and time’ as the latter tends to promote the idea
of thinking of space and time as different things.
1
LECTURE 1. INTRODUCTION & MOTIVATION 2
Recall that one of the weird and wonderful things about QM is that it is not always
possible to simultaneously measure two observables. Mathematically we write this as the
statement that their commutator is non-vanishing:
[ÔA , ÔB ] 6= 0
for non-compatible operators. This essentially tells us that the action of OA on our state
speaks to OB and alters its action and visa versa. However if we are going to preserve
causality, it is important that two spacelike separated operators must not speak to each
other. In terms of our commutator relations, this is2
for any two operators OA and OB . To emphasise, the above condition holds even if OA and
OB are non-compatible.
The collection of observables form a quantum field in spacetime and signals propagate as
disturbances in this field. We interpret these disturbances as particles.3
E ≈ |p|c (1.4)
Now we have a problem! Imagine constraining the particle into a box of size L = ~/2mc
then Equation (1.3) tells us that the uncertainty in the momentum is
∆p ≥ mc,
2
We are working with a field theorist signature of (+, −, −, −).
3
On a technical note, the particle interpretation in QFT is actually much more subtle. A global notion of
particle requires a global notion of time. As we are working in Minkowski spacetime we of course have a global
time coordinate and so this is not a problem. However for a curved spacetime things become more subtle and
you get some highly non-trivial results, namely that black holes have temperature; Hawking Radiation.
4
For a lot more info on this see Simon Rea and my joint notes on Dr. Frederic Schuller’s Quantum Theory
course.
LECTURE 1. INTRODUCTION & MOTIVATION 3
∆E ≥ mc2 .
where H is our ‘one particle state’. To make the above seem a little less abstract, the two
particle state is an element of the space H⊗2 := H ⊗ H, and so each n above corresponds to
allowing n-particle states. The ⊕ symbol just means we need to ‘sum over’ these different
combinations to account for the fact that our system can transition from a n-particle state to
a (n + 1)-particle state.5
Remark 1.1.2 . Anyone familiar with the basic idea of QFT will know that what we never
really just create a single particle in this way, but instead they always appear is so-called pair
production. This is essentially just the statement that a particle is always produced with its
antiparticle, but we’ll return to this later.
QFT will allow us to study how such particle creation processes occur and interpret them
in terms of interacting field. Being a quantum theory, what we will obtain is the probability
that such a process occurs, known as cross sections.
"Quantum fields are the basic ingredients of the Universe, particles are just bundles of
energy and momentum made out of them."
So QFT theory is, as the name suggests, a study of fields, not particles. This might seem
like a pedantic point to make, however this switch in mentality is actually very important for
a solid understanding of the material that is about to come.
5
For more info on the symbols used, see either the QM notes mentioned in footnote 4 or any decent linear
algebra textbook.
LECTURE 1. INTRODUCTION & MOTIVATION 4
1.2.1 Units
Experienced physicists will know that physical units (or dimensions) are important, and must
be kept track of. As you may be aware, it turns out that all physical units can be characterised
by seven others, known as SI units. The three most commonly encountered (and the only
ones used in this course) are mass, length and time. We write the dimensions of a quantity
by square brackets and use the letters [M ], [L] and [T ] for the three units, respectively.
The constants that appear in our equations will, of course, sometimes carry units. The
three dimensionful constants in particle physics are: the reduced Planck’s constant, ~; the
speed of light c; and Newton’s gravitational constant GN . These have dimensions
[L] [M ][L]2 [L]3
[c] = , [~] = , and [GN ] = .
[T ] [T ] [M ][T ]2
It is the former two that will show up a lot (as they already have above) and particle physicists
decided one day they had had enough of writing them out and so decided to switch to a
different set of units, known as natural units. These are defined by setting ~ = 1 = c. From
the above relation this results in
[L] = [T ] = [M ]−1 .
We can therefore just characterise everything in terms of one of these three units. It is standard
to give this privilege to mass, and we obtain mass dimensions. These are just numbers, and
correspond to the power of [M ] appearing. For example
Now recall Equation (1.3). This tells us that position (which has length dimension) has
the opposite dimension to 3-momentum. Then Equation (1.4) tells us that energy has the
same dimension as 3-momentum. We therefore have
[E] = 1. (1.6)
This result is also trivially obtained from E = m (as c = 1). We can therefore express these
dimensions in units of energy. Due to the energy scales that appear in particle physics, we
do not choose to use the Joule but instead use the much more convenient electron volt, eV.
These often appear with the prefixes M, G, T for ‘mega’, ‘tera’ and ‘giga’, respectively.
Example 1.2.1 . As an example, the mass of an electron in these units is
me = 0.511M eV.
1.2.2 Range
Perhaps an obvious question to now ask is "what energy scale does QFT apply to?"
We use the name ‘generalised’ coordinates, because, as we have seen, we don’t need the
same number of generalised coordinates as we do ‘real’ coordinates (i.e. those used to define a
coordinate system). Note that the configuration space is given by the span of the generalised
coordinates.
Example 1.3.1 . The generalised coordinates for the free particle are (x, y), whereas for the
constrained particle we just have one, θ.
Example 1.3.2 . The values (q1 , q2 , q3 ) could be the generalised coordinates for either a single
particle in 3-dimensions or for 3 particles in 1-dimension. This example highlights that in order
to differentiate two systems, we need to know how the configuration space is constructed.
In classical physics, the dynamics of a particle is given by a phase space where qa and pi
are the position and momentum of the particle. For non-accelerating particles, we know that
we can simply specify the position and momentum of a particle at some point and from that
extract out its evolution. In other words, each point in phase space only has one path going
through it, and so by specifying the state of a system we know its evolution.
The question is "how do we find this evolution?" The answer comes from the principle of
least action. These tell us the equations of motion for a system (i.e. how it evolves through
time) and so will give us exactly what we want.
The procedure is as follows: we obtain the Lagrangian L for our system in terms of q and
q̇, integrate it from initial to final time to obtain the action, then we extremise this action
under the constraint that the start and end points remain fixed, and obtain the shortest path
between those points. The calculation is as follows:
Z tf
δS = dtδL qa (t), q̇a (t)
ti
Z tf
∂L ∂L
= dt δqa + δ q̇a
ti ∂qa ∂ q̇a
Z tf tf
∂L d ∂L ∂L
= dt − δqa + δqa
ti ∂qa dt ∂ q̇a ∂ q̇a ti
Z tf
∂L d ∂L
= dt − δqa ,
ti ∂qa dt ∂ q̇a
Example 1.3.3 . Consider the motion of single particle in 1-dimension with a potential U (x).
The Lagrangian is
1
L(x, ẋ) = mẋ2 − U (x),
2
and so out Euler-Lagrange equations are:
mẍ = −U 0 (x),
The above example was to do with a single particle, but what we’re trying to study here are
field theories. We therefore need to adapt the above procedure to this case.
The first change we need to make is to the generalised coordinates. For the particle case,
a was an index which labelled the different q(t)s. However in field theory, we must also
introduce a label for the position spatial of the field, and so our generalised coordinates are
replaced by
qa (t) −→ φ(x) := φa (~x, t),
where ~x is a treated as a label. Note that for the field, we are dealing with a system with an
infinite number of degrees of freedom: one for each value of ~x.
We can also show by considering a chain of connected springs (Do this later from his
notes)6 that we must also replace the Lagrangian with a Lagrangian density, L, related by
Z
L(t) = d3 ~x L φa (x), ∂µ φa (x) . (1.8)
The symmetries of the physics are inserted at the Lagrangian level, by which we mean if
we want a theory that is spatial-rotationally invariant then we require
R(ϕx , ϕy , ϕz )L = L,
(ii) Local Lagrangians; i.e. all of our fields will be evaluated at the same event in spacetime,
(iii) Only contain single derivatives of the fields; i.e. we don’t want anything of the form
∂µ ∂ν φ,
(iv) Renormalisable Lagrangians; this is something that will not really be covered explained
in this course, but basically the coefficients in our Lagrangian must have positive di-
mension. This essentially amounts to not having more than d powers of the field, where
d is the dimension of the spacetime.
These four conditions will actually massively reduce the number of possible Lagrangians we
can consider. We now give a brief justification for each:
(i) We are considering relativistic particles and so we really want our theories to be Lorentz
invariant,
6
I plan to come back and fill this in later. Just right now don’t think it’s worth the time when I have so
many other courses to be focusing on.
LECTURE 1. INTRODUCTION & MOTIVATION 8
(iii) There are two reasons: one is that we would need to alter our Euler-Lagrange equations
to account for higher derivatives, and also adapt our phase space7 ; secondly because we
are fundamentally interested in 4-dimensional spacetime, a Lagrangian that included a
∂µ ∂ν φ term would be non-renormalisable. This comes from the fact that the action has
vanishing mass dimension8 and so, in order to cancel the [dd x] = −d9 contribution from
the integral we require [L] = d. Now if we want to study kinematics, we need a kinetic
energy term in the Lagrangian, this takes the form
but the derivatives have dimension [∂µ ] = 110 and so we get 4 = 2 + 2[φ], or [φ] = 1.
If we were then to include a second order derivative term, Lorentz invariance would
require it appears as
L = a(∂µ ∂ν φ)(∂ µ ∂ ν φ),
but this would then require [a] = −2, which gives a non-renormalisable theory.
(iv) We don’t need to understand this here, but basically it causes a huge pain in perturbation
theory.
Notation. As we have just seen in the explanation of (iii) above, Lorentz invariance requires
that any derivative term comes with the corresponding ‘raised’ derivative term. It is therefore
very useful to introduce a common notion to simply the expressions. We define
This notation can appear a bit confusion because it appears to break Einstein summation
convention (the left-hand side seems to have a µ index but on the right-hand side it is summed
over). For this reason some people (and I am one of these people) prefer to define the notation
(∂φ)2 := ∂µ φ∂ µ φ. (1.10)
This is the notation that will be used in these notes from now on, unless some potential
confusion might arise, in which case we shall write the full expression (i.e. like the right-hand
side above). Similarly we shall use
∂ 2 := ∂µ ∂ µ . (1.11)
Example 1.4.1 . One of the simplest Lagrangians that meets the above conditions is
1 1
L = (∂φ)2 − m2 φ2
2 2 (1.12)
1 2 1 2 1
= φ̇ − ∇ φ − m2 φ2 ,
2 2 2
7
Basically you get equations of motion that contain second order derivatives, and so it is not enough to
just specify q and q̇ on an initial time surface, but we must also specify q̈.
8
Basically because S ∼ ~.
9
Note that integration increases the dimension and so has units of length.
10
They take away length dimension.
LECTURE 1. INTRODUCTION & MOTIVATION 9
where φ = φ(x) is a real scalar field. The Euler-Lagrange equations for this are simply
(∂ 2 − m2 )φ = 0.
This is known as the Klein-Gordan equation. It is very important to note here that this is
the classical Klein-Gordan equation, not the quantum version. As we will see later it has
an quantum analogue (which looks exactly the same) which gives the equations of motion of
spineless particles, i.e. quantum scalars.
Example 1.4.2 . The Klein-Gordan equation above takes the form of a wave equation (i.e. it is
second order in both temporal and spatial derivatives). This is nice,but we know in QM that
the wave equation is replaced by the Schrödinger equation, which is linear in time derivatives
and quadratic is spatial derivatives. We can get a classical analogue to this by considering a
complex scalar field ψ. The Lagrangian is
i
L = (ψ ∗ ψ̇ − ψ̇ ∗ ψ) − ∇ψ ∗ · ∇ψ − mψ ∗ ψ.
2
You treat ψ and ψ ∗ as independant fields. For the ψ ∗ field we have,
∂L i ∂L i ∂L
∗
= ψ̇ − mψ, = − ψ, and = −∇ψ,
∂ψ 2 ∂ ψ̇ ∗ 2 ∂∇ψ ∗
xµ −→ x
eµ = Λµ ν xν
Λµ τ η τ σ Λν σ = η µν . (2.1)
Lorentz transformations are essentially spatial rotations and boosts, and we can work out
the entries of the matrix Λµ ν by consider the action on the {xµ }. We give an example of each
below.
Example 2.1.1 . Let’s set (x0 , x1 , x2 , x3 ) = (t, x, y, z). First let’s consider a rotation around
the z axis. This does nothing to the t axis and obviously doesn’t effect z. We therefore just
set the SO(2) rotation in the xy-plane by angle θ. In other words we have
t t
−→ x cos θ − y sin θ ,
x
y y sin θ + x cos θ
z z
10
LECTURE 2. SYMMETRIES & HAMILTONIAN FIELD THEORY 11
Example 2.1.2 . Following a similar idea to the previous example, we can show that for a boost
along the x axis we have
γ −γv 0 0
−γv γ 0 0
(Λµ ν ) =
0
,
0 1 0
0 0 0 1
√
where γ = 1/ 1 − v 2 .
= φ Λ−1 x . (2.2)
φ(x) −→ φ(x)
e
The above equation is actually a particular case for the action of the representation of the
Lorentz transformation on fields. In this case we are acting on a scalar field and so no D[Λ]
3
Not to be confused with this.
LECTURE 2. SYMMETRIES & HAMILTONIAN FIELD THEORY 12
∂µ j µ = 0. (2.3)
As we will see, symmetries and their conserved currents are incredibly important in field
theory, and so Noether’s theorem is an invaluable tool to field theorists.
L −→ L + ∂µ F µ , (2.4)
δL = ∂µ F µ ,
and so taking these two results away from each other gives us
∂L
jµ = δφa − F µ (φa ) =⇒ ∂µ j µ = 0. (2.5)
∂(∂µ φa )
So this is our conserved current. We get one of these for each continuous δφ transformation
and so we have proved Noether’s theorem.
4
We use the fact that is a constant to take it outside the derivative.
LECTURE 2. SYMMETRIES & HAMILTONIAN FIELD THEORY 13
Before doing some examples, first there is an important comment to make. As we have
been careful to say above, Noether’s theorem tells us that we get a conserved current. This
is a stronger result then saying we get a conserved charge. Indeed a conserved current gives
rise to a conserved charge, and we get this charge as follows:
∂j 0
∂µ j µ = 0 =⇒ = −∇ · ~j.
∂t
Integrating both sides of this over all of space, we have
Z Z
∂ 3 0
d ~x j = − d3 ~x ∇ · ~j
∂t R3 R 3
I
= − d~s · ~j
= 0,
where the last line follows from the fact that we assume ~j falls off sufficiently fast as |~x| → ∞.
We can therefore define our conserved charge by
Z
Q := d3 ~x j 0 . (2.6)
R3
Remark 2.2.2 . Some people also refer to the energy-momentum tensor as the stress-energy
tensor or the stress-momentum tensor, or even the stress-energy-momentum tensor. I will try
to use energy-momentum tensor everywhere here, but I might sometimes just call it the stress
tensor. Apologies for this in advance.
Ok, let’s derive this lovely chappy.
Consider a continuous translation of the spacetime coordinates. As above, we shall work
infinitesimally, and so we have
xµ −→ xµ − µ ,
where we use a minus sign so that the transformation of the field is positive, see the discussion
of active vs. passive transformations above. We can Taylor expand the field to obtain the
transformation
φa (x) −→ φa (x) + µ ∂µ φa (x),
and similarly the Lagrangian transforms as
L(x) −→ L(x) + µ ∂µ L(x).
Comparing this to Equation (2.4), we see that F µ = µ L. Our conserved current is therefore
∂L
jµ = ν ∂ν φa − µ L
∂(∂µ φa )
ν ∂L µ
= ∂ν φa − δν L ,
∂(∂µ φa )
and so we have a conserved current for each ν , i.e. for the translations in each direction.
These are the components of our energy-momentum tensor.
∂L
T µ ν := (j µ )ν = ∂ν φa − δνµ L. (2.7)
∂(∂µ φa )
to be the total energy and the (i-th component of the) total momentum of the field configu-
ration, respectively.
Remark 2.2.3 . Note in deriving Equation (2.7), we didn’t say anything about the actual form
of our Lagrangian, and so this result holds for generic L.
5
Note we have raised the ν index here. As we are working in Minkowksi spacetime this really isn’t a big
deal, however as these are definitions you should be careful in more general spacetimes as some non-trivial
factors will appear.
LECTURE 2. SYMMETRIES & HAMILTONIAN FIELD THEORY 15
ω µν = −ω νµ . (2.11)
Similarly we obtain
δL = −ω µ ν xν ∂µ L.
Now we use a clever trick. Firstly we note that ω µ ν is a constant so we can take it inside the
derivative. Next we note that
∂µ xν = δµν ,
along with ω µ µ = 0 by Equation (2.11). We can therefore also take the xν inside the derivative
and obtain
δL = −∂µ ω µ ν xν L .
This is now of the form Equation (2.4), with F µ = ω µ ν xν L. So our conserved currents are
∂L
jµ = − ω σ ν xν ∂σ φ + ω µ ν xν L
∂(∂µ φ)
∂L
= −ω σ ν ∂σ φ + δ µ σ L xν
∂(∂µ φ)
= −ω σ ν T µ σ xν .
As above we can split this into the individual currents, one for each ω σ ν , and obtain
Each one of these has a corresponding conserved charge. We get the particle result, i.e.
angular momentum, when ρ, σ = 1, 2, 3:
Z
Q = d3 ~x xi T 0j − xj T 0i .
ij
LECTURE 2. SYMMETRIES & HAMILTONIAN FIELD THEORY 16
The question is "what charge do the boosts give us?" Well these correspond to ρ = 0, σ =
1, 2, 3, which give Z
Q = d3 ~x x0 T 0i − xi T 00 .
0i
What is this? Well we know that its temporal derivative vanishes, so using x0 = t, we have
dQ0i ∂T 0i
Z Z Z
3 0i 3 d
= d ~x T + d ~x t − d3 ~x xi T 00
dt ∂t dt
dP i
Z
i d
0=P + − d3 ~x xi T 00 ,
dt dt
where we’ve used Equation (2.9). But we’ve already seen that P i is a constant and so we
simply get Z
d
d3 ~x xi T 00 = constant,
dt
which is the statement that the center of energy of the field travels with constant velocity.
Remark 2.2.4 . Note that Equation (2.11) agrees with the number of Lorentz transformations.
That is, it is an antisymmetric, 4 × 4 matrix and so has 4 × 3/2 = 6 independant entries.
So it has 6 basis elements, 3 of these correspond to spatial rotations and the other 3 are our
boosts.
This is a continuous transformation and so we can work infinitesimally and find the conserved
current. We have
φ01 ≈ φ1 + θφ2 =⇒ δφ1 = θφ2
φ01 ≈ −θφ1 + φ2 =⇒ δφ2 = −θφ1 .
φ2
φ02
θ φ01
θ
φ1
We can check that this obeys ∂µ j µ = 0. First we need the Euler-Lagrange equations for our
action. We have
∂L ∂L
0 = ∂µ − = (∂ 2 + m2 )φ1
∂(∂µ φ1 ) ∂φ1
and similarly for φ2 . Now take the derivative of our current:
∂µ j µ = (∂ 2 φ1 )φ2 + ∂ µ φ1 ∂µ φ2 − (∂ 2 φ2 )φ1 − ∂ µ φ2 ∂µ φ1
= m2 φ1 φ2 − m2 φ2 φ1
= 0,
where we used the Euler-Lagrange equations to get to the third line, and then used the fact
that we have real scalars so φ1 φ2 = φ2 φ1 .
This symmetry is known as a global SO(2) symmetry. The name makes sense: its global
(i.e. the whole plane is rotated) and its a rotation in 2-dimensions. It is actually a specific
case of the more general global SO(N ) symmetry, which has
N (N − 1)
2
conserved currents for N fields.
Example 2.2.6 . We can reformulate the previous example by considering the complex scalar
field
1
ψ = √ φ1 + iφ2 .
2
LECTURE 2. SYMMETRIES & HAMILTONIAN FIELD THEORY 18
L = ∂µ ψ ∗ ∂ µ ψ − m2 ψ ∗ ψ. (2.13)
We are now rotating in the complex plane so instead of considering a SO(2) rotation we
consider the complex U(1) rotation
ψ −→ eiα ψ.
We already know that this is a symmetry because its exactly the same as the previous example,
but what does the current look like in this complexified case? Well simple calculation will
give
j µ = i (∂ µ ψ ∗ )ψ − ψ ∗ (∂ µ ψ) (2.14)
Exercise
Derive the above conserved current. Hint: Note that ψ ∗ −→ e−iα ψ ∗ and then work
infinitesimally.
Remark 2.2.7 . Maybe put a comment here in line with the non-abelian comment made in
Tong, page 18.
Internal symmetries will prove to be vital for the study of particle physics as a QFT. We
will see that the charges arising from such symmetries correspond to things such as electric
charge and particle number.
Above we have constructed classical field theory in terms of Lagrangians. It is true that we
can extend this Lagrangian approach to QFT by using so-called path integrals, and there are
advantages to doing that (most notably that the Lorentz invariance is manifest throughout),
however in this course we shall use the Hamiltonian approach instead. This uses so-called
canonical quantisation to promote the fields to operators.
In particle mechanics, we define the Hamiltonian to be
X
H(p, q) = pa q̇a − L(q, q̇),
a
where
∂L
pa :=
∂ q̇a
is known as the conjugate momentum. The idea is to eliminate q̇ wherever we can and replace
it with p.
In field theory we do a very similar thing, but now we have to use the Lagrangian density,
L, and we define the conjugate momentum density and Hamiltonian density
6
Note that we don’t√ have a factor of 1/2 in this expression as was the case for the real scalar field. This is
just because of the 1/ 2 factor above.
LECTURE 2. SYMMETRIES & HAMILTONIAN FIELD THEORY 19
∂L
π(x) := (2.15)
∂ φ̇(x)
H = π(x)φ̇(x) − L(x). (2.16)
Again we favour replacing any φ̇(x) dependence with π(x) once we know the relation. As
with Equation (1.8), we define the Hamiltonian as the spatial integral over the Hamiltonian
density, Z
H(t) = d3 ~x H(x). (2.17)
Note in this line we appear to have broken Lorentz symmetry as we have picked a preferred
time to define our Hamiltonian. However, as long as we are careful not to do anything silly,
we should be alright because we started with a Lorentz invariant theory. This is what we
mean by the Lorentz symmetry not being manifest: it is not obvious just by looking at the
equations that we have Lorentz symmetry, in contrast to the Lagrangian formalism where
spacetime indices were always summed over and so it was clear.
Example 2.3.1 . As an example let’s consider the real scalar field mentioned above with La-
grangian7
1 1
L = (∂φa )2 − m2 φ2a
2 2
1 2 1 1
= φ̇a − (∇φa )2 − m2 φ2a .
2 2 2
The conjugate momentum is therefore
H = φ̇2a − L
1 1 1 2 2
= φ̇2a + (∇φa )2 + m φa
2 2 2
1 2 1 1 2 2
= πa + (∇φa )2 + m φa ,
2 2 2
where in the last line we have done our procedure of swapping out the φ̇s for πs using the
relation above.8 The integral over these terms give three contributions to the Hamiltonian,
they are Z
3 1 2 1 2 1 2
H = d ~x π + (∇φa ) + φa .
2 a |2 {z } |{z}
|{z} 2
kinetic shear mass
7
It is incredibly common in the field theory world to forget to say ‘density’, as I have done here (and
probably above too). The symbols should tell you what we mean: densities are normally given in fancy curly
font.
8
For clarity, you don’t just swap φ̇ −→ π, but you use the relation for π in terms of φ̇ to eliminate the φ̇s.
In this particular case, π = φ̇, and it does correspond to just swapping them.
3 | Canonical Quantisation & Free Klein
Gordan Field
Recall the idea behind quantisation in QM is to promote the generalised coordinates in the
classical theory and ‘promote’ them to operators acting on the Hilbert space of the quantum
theory. This recipe is called canonical quantisation. The Poisson bracket1 relation between
the generalised coordinates magically transformed into a commutation relation between the
operators. That is2
{qa , ab } = {pa , pb } = 0 −→ [q̂a , q̂b ] = [p̂a , p̂b ] = 0
{qa , pb } = δab −→ [q̂a , p̂b ] = iδab
As we are working in Minkowski spacetime, we can lower the indices on the ps easily and
obtain
[q̂a , q̂b ] = [p̂a , p̂b ] = 0 and [q̂a , p̂b ] = iδab .
We adopt the same philosophy for fields, and promote them to operator valued functions.
However we have a small problem: for the particle mechanics case we just had a finite number
of generalised coordinates and so it was easy to do, whereas for the fields there’s an infinite
number, one for each point ~x in space. We treat this spatial dependence as a label and so
need something analogous to the δab above. The answer is obviously the usual delta function.
We use the Schrödinger picture, where all time dependence appears in the states |ψi which
obey the Schrödinger equation
d |ψi
i = H |ψi ,
dt
and the operators themselves are time independent, and write3
[φa (~x), φb (~y )] = [πa (~x), πb (~y )] = 0, and [φa (~x), πb (~y )] = iδ (3) (~x − ~y )δab . (3.1)
Remark 3.1.1 . Note we have really butchered our manifest Lorentz invariance here: we com-
pletely separated space and time and made the operators only functions of space! Of course
it must be true that we’re alright to do this and still get a Lorentz invariant theory, otherwise
we wouldn’t be doing it in these notes. However this choice of doing things will introduce
some factors here and there (e.g. to the measures in integrals) to ensure Lorentz invariance.
1
Note to self: maybe put something about Poisson brackets above when discussing generalised coordinates.
2
There are factors of ~s in these equations, but in natural units they go bye-bye.
3
You can work in the Heisenberg picture and define what are known as ‘equal time’ commutation relations.
20
LECTURE 3. CANONICAL QUANTISATION & FREE KLEIN GORDAN FIELD 21
where to get to the second line we have identified φe∗ (−~ e p, t). Using this coordinate
p, t) = φ(~
system, our Klein-Gordan equation reduces to (dropping the tilde and using the argument to
indicate which function it is)
2
∂ 2
+ (~
p + m ) φ(~ p, t) = 0.
∂t2
Then the keen-eyed person notes that for each value of p~ this corresponds to its own
harmonic oscillator with frequency
p
ωp~ = p~2 + m2 . (3.2)
To stress the point, we get a harmonic oscillator each every single value of p~ independently
from any other value. The general field φ(~x, t) is then simply given by a linear superposition of
(an infinite number of) harmonic oscillators. We therefore have, at least classically, achieved
the goal above to frame the theory in such a way that each degree of freedom (each p~) evolves
independently from the others. Now what we want to do is quantise it.
4
Again this comment is just made because things are different in general spacetimes, see footnote 3 from
the first lecture.
LECTURE 3. CANONICAL QUANTISATION & FREE KLEIN GORDAN FIELD 22
Remark 3.2.2 . Note that in order to take the Fourier expansion above we need to assume
that the field φ(~x, t) die off sufficiently quickly as |~x| → ∞. In order to guarantee this, we
use fields from the so-called Schwartz space. A precise definition (with a lot more context for
why they’re useful in QM) can be found in Simon Rea and my notes on Dr. Schuller’s course
on quantum theory, but for here we shall just say they are functions that die off at infinity
as do their derivatives.
[q, p] = i. (3.4)
We can rewrite this Hamiltonian in a nicer form by introducing the creation and annihilation
operators: a† and a, respectively. These are defined such that
r
1 † ω
a + a , and p = −i a − a† . (3.5)
q= √
2ω 2
Exercise
Show that Equation (3.4) and Equation (3.5) give
[a, a† ] = 1. (3.6)
Then using this result show that the Hamiltonian Equation (3.3) can be rewritten as
† 1
H =ω a a+ . (3.7)
2
We can now begin to look at the spectrum of the QHO. The ground state |0i is defined
via
a |0i = 0 (3.9)
and so we see from Equation (3.7) that the ground state has energy
ω
E0 = .
2
LECTURE 3. CANONICAL QUANTISATION & FREE KLEIN GORDAN FIELD 23
As the notation suggests (and as can easily be seen from Equation (3.5)) the creation and
annihilation operators are Hermitian conjugates to each other. This translates into bra-ket
notation in terms of left and right actions:
†
a |ψi = hψ| a† . (3.12)
d3 p~
Z
1 h p·~
i~ x † −i~ p·~
x
i
φ(~x) = a p
~ e + ap~ e
(2π)3 2ωp~
p
(3.13)
d3 p~
r
ωp~ h
Z i
p·~
i~ x † −i~ p·~
x
π(~x) = (−i) ap~ e − ap~ e ,
(2π)3 2
where the form of these expressions makes sense when comparing to Equation (3.5).
Remark 3.2.3 . Note that it is only at this point that we are now considering the quantum
Klein-Gordan field. This is in contrast why I was careful to emphasise before that we were
considering the classical Klein-Gordan field. The equations of motion are still the same, but
now they are quantum expressions.
5
We’re ignoring normalisation here, i.e. hn|ni 6= 1. This is not important for our present discussion.
LECTURE 3. CANONICAL QUANTISATION & FREE KLEIN GORDAN FIELD 24
This claim is not complicated to prove, but a bit of a pain to write out and so I, lovingly,
decided to set them as an exercise.
Exercise
Prove Claim 3.2.4. Note it is an if and only if statement so you need to show it both
ways, i.e. the φ/π commutators imply the a/a† ones and visa versa. Hint: If you get
very stuck, Prof. Tong sketches one on page 24 of his notes.
Remark 3.2.5 . Equation (3.14) seem strange: the left-hand side is a commutator between
operators whereas the right-hand side is a delta distribution? The way we wrap our heads
around this is to remember that any physical results in QM appear in inner products hψ| A |ψi,
which can be written as integrals and so the delta function makes sense.
Using Equation (3.12), and the fact that our states are orthogonal, we see that it is only
the terms that contain ap~ ap†~ or ap†~ ap~ that contribute to expectation values. We will see an
example of this below when finding the momentum of the first excited state.
This idea, and similar ones where we get C-numbers on the right-hand side, comes up
again and again in QFT. What we have to keep telling ourselves is that "remember we’re
sandwiching this between states!" and then go from there.
Ok, so we have already derived that the Hamiltonian for this system is
Z
1
d3 ~x π 2 + (∇φ)2 + m2 φ2 .
H=
2
d3 p~ ωp~ h
Z i
† †
H= ap~ ap~ + ap~ ap~
(2π)3 2
Z 3
(3.15)
d p~ † 1 3 (3)
= ωp~ ap~ ap~ + (2π) δ (0) ,
(2π)3 2
Remark 3.2.6 . All jokes aside about the above exercises, it is actually a really beneficial
exercise to do and will test your understanding of a reasonable amount of the information
leading up to here. So honestly at least give them a go.
LECTURE 3. CANONICAL QUANTISATION & FREE KLEIN GORDAN FIELD 25
ap~ |0i = 0 ∀~
p. (3.16)
We can then use Equation (3.15) to find the ground state energy:
Z
3 1 (3)
H |0i = d p~ ωp~ δ (0) |0i ,
2
which is... infinite?! Uh oh this does not seem good at all, but what did we do wrong? The
answer is actually nothing. QFT is filled with infinities and, as Prof. Tong explains "each
tells us something important, usually that we’re doing something wrong, or asking the wrong
question". What we need to do when infinities arise is ask where they come from and what
that implies. So let’s investigate this one.
Firstly the bad news... it’s actually two infinities! The first one comes from the fact that
we’re considering the entirety of R3 . So instead let’s put our theory inside a box of size L.
We adopt the usual idea where we prescribe periodic boundary conditions to the sides, and
so get the flat torus.6 Finally we just let L → ∞ to get our result back. We can then use
Z L/2 Z L/2
3 (3) 3 x·~
i~ p
d3 ~x = V,
(2π) δ (0) = lim d ~x e = lim
L→∞ −L/2 p
~=0 L→∞ −L/2
where V is the volume of our theory. Ok so the (2π)3 δ(0) infinity is just because we’re
considering the total energy E0 instead of the energy density
d3 p~ 1
Z
E0
ε0 := = ωp~ .
V (2π)3 2
This is still infinite though! Why? well because we take the positive root in Equation (3.2)
and so we’re summing an infinite number of positive numbers! This seems like a more tricky
beast to tame, however then a light bulb goes off in our heads and we realise that this is
just the ground state energy, and so, from the extension of Equation (3.11), it will appear in
all the energies. Then remembering that all we can measure physically is energy differences
we realise that we are free to just ‘take this away’ from every energy in our system (as the
difference wont be effected), so this is what we shall do.
You might not be very comfortable with ‘subtracting infinity’, and if that is the case,
allow me to provide a calming remark: we are mere mortals pretending to be Gods. In less
poetic (and perhaps offensive) words: we have assumed that the theory we have written down
is valid up to arbitrarily large momentum. This corresponds to arbitrarily large energies,
or arbitrarily low length scales. Nature is highly unlikely to agree with us and so plays the
trump card of "you need to cut-off your theory at high momenta!" If we do this, the ground
state energy density will become finite and then we can comfortably take it away from all
energies and proceed as if nothing ever happened. Infinities of this kind are called ultra-violet
divergences.
6
If you’re confused why I say flat torus, either google it or think about what happens geometrically when
you make these periodic boundary conditions.
LECTURE 3. CANONICAL QUANTISATION & FREE KLEIN GORDAN FIELD 26
[H, ap~ ] |0i = −ωp~ ap~ |0i , and [H, ap†~ ] = ωp~ ap†~ |0i , (3.17)
H |~
p, ~q, ...i = (ωp~ + ωq~ + ...) |0i .
Remark 3.2.7 . Just as with the states |ni defined in Equation (3.10), our states |~
pi are not
normalised. They are orthogonal though. We shall return to the normalisation later.
and have already remarked that this is the energy of a relativistic particle. But could this
just be some coincidence? Well let’s look at the physical momentum given by Equation (2.9).
Straight forward calculation using the Lagrangian for our system gives
T µν = ∂ µ φ∂ ν φ − η µν L,
7
And perhaps it has exhausted you getting to this point.
8
Recall that we’re using the field theorist’s convention of (+, −, −, −).
LECTURE 3. CANONICAL QUANTISATION & FREE KLEIN GORDAN FIELD 27
d3 p~ p~ h †
Z i
~ † † †
P = a a p
~ + ap~ a + ap~ a ~
−p + a a
(2π)3 2 p~ p~ p~ −p
~
Z 3
d p~
= p~a† ap~ ,
(2π)3 p~
where to get to the last line we have used the commutation relation between a/a† and then
dropped the δ(0) term as we did for the energy, and then ‘dropped’ the aa and a† a† terms
using the argument of Remark 3.2.5. So we see that the first excited state has momentum
P~ |~
pi = p~ |~
pi ,
which is exactly what we expect for a particle. We therefore interpret the state |~
pi to be a
particle with 3-momentum p~. It is important to note that this is simply an interpretation,
what we’re really dealing with is an excitation of a field.
Exercise
Using the definition Z
i ijk
J = d3 ~x (J 0 )jk ,
where ijk is the Levi-Civita tensor density and J is the classical angular momentum
defined in Equation (2.12), show that the state |~
p = 0i has
J i |~
p = 0i = 0.
The result of the above exercise tells us that the our interpreted particle has no spin (i.e.
no internal angular momentum). That is, our quantisation of the Klein-Gordan field gives
rise to a spin-0 particle.
As we have used n creation operators, we refer to these states are n-particle states and
interpret it as n particles with 3-momenta (~ p1 , ..., p~n ). As the creation operators commute
with each other we have
|~
p, ~qi = |~q, p~i ,
and so the particles are symmetric under interchange. We therefore see that these particles
are bosons. This also agrees with the fact that they are spin-0, as per the above exercise.
Now recall that in the first lecture we showed when we unite special relativity with QM
we have to account for particle number changing. We said that we do this by changing our
LECTURE 3. CANONICAL QUANTISATION & FREE KLEIN GORDAN FIELD 28
Hilbert space to be a Fock space, Equation (1.5). We can clarify a bit what we meant there
now. If H is the Hilbert space of our 1-particle states (i.e. |~
pi ∈ H), then we have
We then take the direct sum over all these different states so that our total Hilbert space
accounts for the particle creation process. That is, let’s say we start off with a state |ψi ∈ H⊗n
and then something happens to it and causes the production of a new particle. The new state
would then be an element of H⊗(n+1) . If we don’t take the Fock space construction as above,
this process would mean that we leave our Hilbert space, which is a big no-no.
The next thing we can ask is how does the action of φ(~x) fit into our particle interpretation?
Well let’s calculated it and see. We have
d3 p~
Z
1 h p·~
i~ x † −i~p·~
x
i
φ(~x) |0i = ap~ e + ap~ e |0i
(2π)3 2ωp~
p
d3 p~
Z
1
= 3
p e−i~p·~x |~
pi .
(2π) 2ωp~
9
R
This result comes from inserting a complete set of states, 1 = dp |pi hp|. See any decent QM book for
details.
LECTURE 3. CANONICAL QUANTISATION & FREE KLEIN GORDAN FIELD 29
This tells us that once we are in a particular H⊗n in the Fock space, we will never leave
it in a free theory. This makes them a bit boring to study. Fortunately later we will allow for
interactions in our theory and see that the above result no longer holds in general. The sad
news is free theories are the only ones we can solve exactly, and so in order to study interacting
theories we are going to have to develop the mathematics of Feynman diagrams. These are
essentially life saving tools to calculate things in perturbation expansions of interacting QFTs.
4 | Restoring Lorentz Invariance & Causal-
ity
Recall that last lecture we gave an interpretation of our results as particles. There is a problem
with this interpretation: they are momentum eigenstates and so, by Heisenberg uncertainty
principle, are not localised in position space at all. This problem essentially stems from the
fact that in QM the position and momentum eigenstates are not good elements of the Hilbert
space as the normalise to delta-functions. The QFT equivalent is that the operators φ(~x)
and ap~ are not good operators on the Fock space as the resulting states are not normalisable.
Explicitly we have
These are known as operator valued distributions, and the way we fix this problem is by
smearing them over space. For example the wavepacket
d3 p~ −i~p·~x
Z
|ϕi = e p) |~
ϕ(~ pi ,
(2π)3
is partially localised in both momentum and position space. A typical state is an element of
the Schwarz space and is of the form
2 /2m2
p) = e−~p
ϕ(~ .
30
LECTURE 4. RESTORING LORENTZ INVARIANCE & CAUSALITY 31
If our expectation value is going to be unchanged, we would need the transformation on the
states to be unitary, i.e.
|~
pi −→ U (Λ) |~
pi ,
but there is absolutely no reason why this would be the case.
So what do we do? We need to normalise our one-particle states so that we somehow get
a Lorentz invariant inner product. We do this using two tricks.
4.2.1 Trick 1
The first thing we note is that the integral d4 p is manifestly Lorentz invariant, provided the
R
integrand is also Lorentz invariant. Then we note that the relativistic dispersion relation
pµ pµ = m2
is also manifestly Lorentz invariant. This can be written as
p20 = p~2 + m2 = ωp~2 ,
which has two branches, p0 = ±ωp~2 , as depicted below.
p0
p~
The choice of branch is Lorentz invariant, so w.l.o.g. we take p0 > 0. We we know the
integral
d4 p
Z
2πδ(p2 − m2 )Θ(p0 )
(2π)4
where Θ is the Heaviside function, is Lorentz invariant. We can rewrite this in polar coorindates
as
Z ∞ Z ∞
d3 p~ d3 p~ d3 p~ 1
Z Z Z
2 2 2 dp0 0
dp 0 δ(p 0 − p
~ − m ) = δ(p − ωp~ ) =
(2π)3 0 (2π)3 0 2p0 (2π)3 2ωp~
where the middle term comes from
1
δ f (x) − f (x0 ) = δ(x − x0 ).
|f 0 (x0 )|
The thing we started with was Lorentz invariant and so we know that that the measure
d3 p~ 1
Z
(4.1)
(2π)3 2ωp~
is a Lorentz invariant measure.
LECTURE 4. RESTORING LORENTZ INVARIANCE & CAUSALITY 32
d3 p~ 1
Z
2ωp~ ap†~ |0i ,
p p
3
h0| 2ωq~ aq~
(2π) 2ωp~
where on the second equality we have just used E, for energy, as this is standard notation,
then our inner products become Lorentz invariant.
As we remarked above, we slaughtered our Lorentz invariance through our quantisation pro-
cedure. Even besides the above mix of the measure, we could have easily seen (or at least
been cautious) by the fact that we are working in the Schrödinger picture, so our operators
φ(~x) only depend on space, and not time. Equally our one states evolve in time, with the
evolution given by the Schrödinger equation, which implies
[φa (~x, t), φb (~y , t)] = [πa (~x, t), πb (~y , t)] = 0, and [φa (~x, t), πb (~y , t)] = iδ (3) (~x − ~y )δab . (4.7)
LECTURE 4. RESTORING LORENTZ INVARIANCE & CAUSALITY 33
We can now study to see if the Heisenberg equations, Equation (4.5), for φ(x) and π(x)
are equivalent to the Klein-Gordan equation. Using
Z
1
H = d3 ~x π 2 + (∇φ)2 + m2 φ2 ,
2
and the commutations above
Z
d ih i
φ= d3 ~y π 2 (y), φ(x)
dt 2
Z
i
= d3 y π(y)[π(y), φ(x)] + [π(y), φ(x)]π(y)
2
Z
= i d3 ~y π(y)(−i)δ (3) (~y − ~x)
= π(x).
That is,
φ̇(x) = π(x). (4.8)
Exercise
Using Equation (4.5) show that
or more familiarly,
(∂ 2 + m2 )φ = 0,
the Klein-Gordan equation! We refer to this as the correspondence principle.
Exercise
Show that
eiHt ap~ e−iHt = e−iEp~ ap~ , and eiHt ap†~ e−iHt = eiEp~ ap†~ .
Hint: Use
∞
(iHt)n
eiHt ap†~ e−iHt ap†~ e−iHt , and [H, ap†~ ] = Ep~ ap†~ .
X
=
n!
n=0
We can find the mode expansion of φ(x) in the Heisenberg picture by starting with the
Schrödinger expression
d3 p~
Z
1 i~p·~x −i~ x †
p·~
φ(~x) = e ap~ + e ap~
(2π)3 2Ep~
p
LECTURE 4. RESTORING LORENTZ INVARIANCE & CAUSALITY 34
and then use Equation (4.4) and the result of the previous exercise to obtain
d3 p~
Z
1 −iEp~ t+i~p·~x x †
p·~
iEp~ t−i~
φ(x) = e ap~ + e ap~
(2π)3 2Ep~
p
(4.10)
d3 p~
Z
1 −ipx
† ipx
= a p
~ e + a p
~ e .
(2π)3 2Ep~
p
p0 =Ep~
You can then quickly arrive at the expression for π(x) by using Equation (4.8), giving
r
d3 p~ Ep~ † ipx
Z
−ipx
π(x) = i ap~ e − ap~ e . (4.11)
(2π)3 2
p0 =Ep~
These look a lot like Equation (3.13), but now in terms of 4-vectors. This seems more
manifestly Lorentz invariant!
So we see that φ(x) is a superposition of plane waves e±ip0 t , but the coefficients create and
destroy particles. We can therefore think of the field as a ‘hammer’ which bashes the vacuum
and shakes quanta out of it. That is the operator acting on the vacuum creates a particle:
d3 p~ 1 −i~p·~x
Z
φ(~x, 0) |0i = e |~
pi , =⇒ p| φ(~x, 0) |0i = e−i~p·~x .
h~
(2π)3 2Ep~
If we act on an n particle state instead, because φ(x) contains both creation and annihilation
operators, it has a amplitude to create both (n + 1) and (n − 1) particle states.
4.4 Causality
As we have mentioned above, in order for our theory to be causal, we want spacelike separated
operators to commute, that is
This statement is obviously made in the Heisenberg picture, as the operators are functions of
spacetime, not just space. So we need to study the commutator between two fields [φ(x), φ(y)],
but where now we don’t impose equal time, i.e. x0 6= y 0 in general. We define
d3 p~ 1 −ip(x−y)
Z
ip(x+y)
= e − e
(2π)3 2Ep~
The first thing we note is that this expression is Lorentz, as we have the correct measure,
Equation (4.1), and the exponential contain 4-vectors. We can therefore choose a specific
reference frame in order to evaluate its timelike/spacelike properties.
LECTURE 4. RESTORING LORENTZ INVARIANCE & CAUSALITY 35
(i) First look at the spacelike separations, i.e. (x − y)2 > 0. Let’s work in a reference frame
where (x − y) = (0, ~a), then we have
d3 p~ 1 i~p·~a
Z
−i~
p·~a
∆(0, ~a) = e − e .
(2π)3 Ep~
Now note that Ep~ = p~2 + m2 and so E−~p = Ep~ , and also that the measure is invariant
p
under p~ −→ −~ p, as you get a minus sign but the integration limits flip. So we can freely
change the sign in the second exponential, and then the two terms cancel, i.e.
∆(0, ~a) = 0.
(ii) Let’s also check that it doesn’t vanish for timelike separations (which would be an
obvious problem). Again can pick a frame, so simply choose (x − y) = (t, 0), then we
have
d3 p~ −iEp~ t
Z Z p
iEp~ t
∼ d3 p~ sinh
∆(t, 0) = 3
e −e p~2 + m2 t ,
(2π)
which does not vanish (its a Bessel function).
Remark 4.4.1 . Note that the objects on the right-hand side of Equation (4.12) are operators.
However we’ve just shown that it comes out to be a C-number. Essentially what we have
done is sandwich it in between two vacuum states, as per Remark 3.2.5. In other words we
really should have written
h0| [φ(x), φ(y)] |0i ,
but its common field theory notation to drop the bra and ket. We will continue to do this next
lecture, but this is just another remark to go with Remark 3.2.5 to remind us that everything
needs to be sandwiched.
5 | Complex Scalar Field & Propaga-
tors
Recall the classical Lagrangian for the complex scalar field, Equation (2.13),
L = ∂µ ψ ∗ ∂ µ ψ − m2 ψ ∗ ψ.
∂ 2 + m2 ψ = 0, and ∂ 2 + m2 ψ ∗ = 0.
When we quantise this theory we again treat each field as an independent degree of freedom.
In quantising we replace the complex conjugate with the Hermitian conjugate. We want to
obtain an expansion equivalent to that of Equation (3.13), but we have to note something:
the field ψ is not real and so the corresponding quantum operator will not be Hermitian.
We therefore need to have different labels for our creation and annihilation operators in the
expansion for ψ. We then get the expansion for ψ † by taking the Hermitian conjugate. The
result is
d3 p~
Z
1 i~p·~x † −i~p·~
x
ψ= bp~ e + cp~ e
(2π)3 2Ep~
p
(5.1)
d3 p~
Z
† 1 p·~
i~ x † −i~p·~
x
ψ = cp~ e + bp~ e
(2π)3 2Ep~
p
Remark 5.1.1 . Note that the dagger comes with the negative exponential. This is because we
want the particles we create to have positive frequency, which corresponds to e−i~p·~x .
Classically we had the result π = ∂L/∂ ψ̇, Equation (2.15). We turn this into a quantum
relation, and using ∂L/∂ ψ̇ = ψ̇ ∗ , we obtain the operator expressions
r
d3 p~ Ep~
Z
p·~
i~ x † −i~p·~
x
π= i − cp~ e + bp~ e
(2π)3 2
r (5.2)
d3 p~ Ep~ i~p·~x
Z
† † −i~
p·~
x
π = (−i) bp~ e − cp~ e
(2π)3 2
36
LECTURE 5. COMPLEX SCALAR FIELD & PROPAGATORS 37
Remark 5.1.2 . I have chosen to write the above equations so that the creation operator always
appears to the far most right. This means that the bs and cs move around, obviously. There
are two reasons I am writing them like this: it’s how Prof. Spannowsky writes them, and it
then they take the same form as Equation (3.13). Obviously it doesn’t matter which order I
write them in, I just make this remark because Prof. Tong writes them so that the b term is
always to the left (see pages 33/34 of his notes).
We want to keep the same form for the equal time commutation relations as for the real
scalar field, i.e.
[ψ(~x), π(~y )] = iδ (3) (~x − ~y ), [ψ(~x), π † (~y )] = 0, (5.3)
and similarly with Hermitian conjugates everywhere,1 and the vanishing ones
Exercise
Show that the above commutation relations hold if, and only if, we also have
Exercise
Derive the following expression for the Hamiltonian,
d3 p~
Z
1
Ep~ bp†~ bp~ + cp†~ cp~ . (5.5)
H= 3
2 (2π)
Hint: Derive the Hamiltonian density and use Equations (5.1) and (5.2). You will need
to drop an infinity term as we did for the real scalar field.
In our particle interpretation, Equation (5.5) tells us that the complex theory contains
two particles of the same energy, and therefore mass. We can show that they also both have
vanishing spin. The states of our Hilbert space are made up of products of these two particle
types, we use a semi-colon in the bra/ket to indicate the different types. For example
The classical complex field conserved current, Equation (2.14), can be upgraded to the
quantum theory and from that we can obtain the conserved charge
d3 p~ †
Z
cp~ cp~ − bp†~ bp~ = Nc − Nb , (5.6)
Q= 3
(2π)
which says that the number of antiparticles minus the number of particles is conserved locally.
Of course in the free theory we have that Nc and Nb are separatly conserved, so this statement
is nothing new. However, as we keep teasing, in the interacting theory the number of each
type of particle will no longer be a conserved charge, but Equation (5.6) will still hold. The
locality of this result will translate into the fact that particles and antiparticles are always
created in pairs, and they can annihilate to produce something different (e.g. a photon).
Remark 5.1.3 . As we just said, these are spin-0 particles and so do not correspond to Fermions
(i.e. electrons etc), but they give us a taste of what’s to come. We can refer to this complex
scalar field as a ‘poor man’s Fermion’.
5.2 Propagators
At the end of the last lecture we checked to see if our theory was causal by checking that the
operators commuted for spacelike separation. We could have worded this question differently,
and perhaps more ‘particle physicsy’. We could have asked "what is the probability for a
particle to propagate from point y to point x?" If x and y are spacelike separated, we would
want the answer to be a big fact zero, otherwise the particles would be travelling faster than
light. The propagation is given by the 2-point function h0| A(x)A(y) |0i where A is our field
in the Heisenberg picture. So we need to consider this calculation theory by theory.
d3 p~ d3 ~q
Z
1 1
= 3 3
p p h0| ap~ a†q~ |0i e−ipx+iqy
(2π) (2π) 2Ep~ 2Eq~
d3 p~ 1 −ip(x−y)
Z
= e ,
(2π)3 2Ep~
h0| ap†~ = 0 = aq~ |0i , and h0| ap~ a†q~ |0i = (2π)3 δ (3) (~
p − ~q).
Again this result is Lorentz invariant and so we can pick any frame we like. We want to check
spacelike separations (x − y)2 < 0, so choose a frame with (x − y) = (0, ~r), and work in polar
LECTURE 5. COMPLEX SCALAR FIELD & PROPAGATORS 39
This is an integral in the complex plane with two poles at p = ±im. We therefore need to
take two branch cuts as indicated in the figure below. We do the integral over the shaded
region and extend out to infinity.
im
−im Re(p)
D(x − y) ∼ e−m|~x−~y| ,
But the above formula has |~x − ~y |, so it doesn’t change when we switch x ←→ y. This is why
the result vanished last lecture. So what does this correspond to in our particle propagation
terms? Well because we’re considering a spacelike path, their is no Lorentz invariant way to
order events, so we have to consider both the propagation x → y and y → x, but these have
the same amplitude, and so cancel.
LECTURE 5. COMPLEX SCALAR FIELD & PROPAGATORS 40
Then using the relevant definitions and commutators show these both give
d3 p~ 1 −ip(x−y)
Z
e
(2π)3 2Ep~
and a similar calculation to the real scalar gives that this vanishes for spacelike separations.
So we give the same particle propagation tale, but now with an interesting difference. As
we see in Equation (5.7), Db (x, y) corresponds to an antiparticle propagating from y → x
whereas Dc (y − x) corresponds to a particle propagating from x → y. So now the cancellation
in amplitude comes from a particle going one way while an antiparticle going the other way!
So causality requires that every particle have a corresponding antiparticle with the same mass
but opposite quantum numbers. This isn’t actually a new interpretation, it’s just that for the
real scalar field the particle is its own antiparticle so we didn’t notice.
where the cases on the right hand side defines T , the time ordering operator:
T O1 (t1 )O2 (t2 ) = O1 (t1 ), O2 (t2 )Θ(t1 − t2 ) + O2 (t2 ), O1 (t1 )Θ(t2 − t1 )
(5.9)
(
O1 (t1 ), O2 (t2 ) t1 > t2
= .
O1 (t2 ), O2 (t1 ) t2 > t1
Physically what this tells us is that first the particle is created and then destroyed. It will
be incredibly useful for us to write this expression as a 4-integral, and this comes in the form
of Feynman’s trick.
Claim 5.2.1 . We can write the Feynman propagator as
LECTURE 5. COMPLEX SCALAR FIELD & PROPAGATORS 41
d4 p
Z
i
∆F (x − y) = e−ip(x−y) , (5.10)
(2π)4 p2 − m2 + i
Proof. We prove this claim by taking a complex contour integral and using Cauchy’s theorem.
First let’s rewrite the right-hand side of Equation (5.10) slightly:
1 1
= 2
p2 − m2 + i p0 − p~2 − m2 + i
1
= 2
p0 − Ep~2 + i
1
= 2
p0 − (Ep~ − i)2
1 1
=
p0 − (Ep~ − i) p0 + (Ep~ − i)
where to get to the penultimate line we used the fact that is infinitesimal, and relabelled
= 2Ep~ . Explicitly, we have
then we drop the 2 term and, because Ep~ > 0, we can redefine → 2Ep~ which is still positive
and infinitesimal.
We now have an integral with poles
p±
0 = ±(Ep
~ − i)
y 0 > x0
p−
0
Re(p0 )
p+
0
x0 > y 0
The shaded regions are meant to indicate how we close our contours for the relevant (x0 − y 0 )
sign. The arrows are meant to indicate which way round we close each contour (so x0 > y 0 is
clockwise and y 0 > x0 is anticlockwise). We pick up the relative poles as indicated.
LECTURE 5. COMPLEX SCALAR FIELD & PROPAGATORS 42
Let’s consider the case x0 > y 0 , then we pick up p+ 0 pole, and using Cauchy’s theorem,
I
f (z)
dz = (2πi)f (z0 ),
(z − z0 )
we get the residue −1/2Ep~ , where the minus sign comes from that fact that we’re doing a
clockwise integral,2 and obtain
d4 d3 p 2πi i −iEp~ (x0 −y0 )+i~p·(~x−~y)
Z Z
i −ip(x−y)
e = e ,
(2π)4 p2 − m2 + i (2π)3 2π 2Ep~
but the right-hand side (after cancelling) is just D(x − y).
Exercise
Finish the proof above. That is show
∆F (x − y) = D(y − x) if y 0 > x0 .
Remark 5.2.2 . If you are not 100% comfortable with the idea of putting the i in the denom-
inator, you can obtain the same result but now the poles like on the real axis. The contour
you need to take to get the Feynman propagator is drawn below.
Im(p0 )
−Ep~
Re(p0 )
+Ep~
Exercise
Show that the diagram in the above remark will lead to the same result for the Feynman
propagator. Hint: If you get stuck, this is how Prof. Tong does it. But I advise you
try it first using the calculation done in the proof above for guidance.
The Feynman propagator is, in fact, also a Green’s function3 for the Klein-Gordan equa-
tion. We see this easily using Equation (5.10) (with = 0, as we don’t need to use a contour
integral here):
d4 p
Z
i
2 2
(∂ + m )∆F (x − y) = (∂ 2 + m2 )e−ip(x−y)
(2π) p − m2
4 2
d4 p
Z
i
− p2 + m2 e−ip(x−y)
= 4 2 2
(2π) p − m
d4 p −ip(x−y)
Z
= −i e
(2π)4
= −iδ (4) (x − y).
2
If this doesn’t make sense, basically in Cauchy’s theorem you use the anticlockwise contour integral.
3
Basically a function that gives a delta function when acted on by a differential operator.
LECTURE 5. COMPLEX SCALAR FIELD & PROPAGATORS 43
This is a really nice result as it allows us to ‘invert’ the Klein-Gordan equation, i.e. turn it
from a differential equation to a integral one.4
As we have seen, in free theories our equations of motion are linear in the fields, e.g. the
Klein-Gordan equation
(∂ 2 + m2 )φ = 0.
We solved these theories by a Fourier analysis and saw that all the different modes decoupled.
That is we never had a ap~ aq~ term in our Hamiltonian etc. This lead to us showing that the
number operator in such theories in conserved, Equation (3.19). This is why we called them
free theories.
Obviously the in the real world things do interact, and we have hinted a few times that
this will lead to the number operator not being conserved anymore. We need some way, then,
to construct interacting theories and the rest of this lecture is dedicated to exactly that.
We get interactions in our theory when the equations of motion are not linear in the fields.
Recalling that the equations of motion come from varying the Lagrangian (i.e. they’re the
Euler-Lagrange equations). So if we want interaction terms in our equations of motion, we’re
going to have to add interaction terms into the Lagrangian:
L = Lfree + Lint .
This will also lead to interaction terms in the Hamiltonian, which we want.
However, in QFT we cannot just arbitrarily change our Lagrangian. That is, we need
to keep our ‘suitable Lagrangian’ conditions we introduced previously. We shall state them
again here, but now in terms of the additional Lint :
(i) We only want local interactions, in order to preserve causality. That is something of the
form Z
d3 yφ(x)φ2 (y)
is not allowed.
(iii) Lint respects the internal symmetries. For example, if our free theory was U(1) invariant
(i.e. ψ → eiα ψ) then we cannot have something like ψψ in Lint .
5.3.1 Examples
Example 5.3.1 . As an example consider the popular φ4 theory. This has Lagrangian
1 1 λ
L = (∂φ)2 − m2 φ2 − φ4 . (5.11)
|2 {z 2 } |4!{z }
Lfree Lint
In terms of Feynman diagrams (to come shortly) the interaction Lagrangian will correspond
to something of the form
p1 p4
p2 p3
Example 5.3.2 . In the above theory we only had one field φ, but we can have (and we will
definitely need) interactions between different fields. As an example, we have the scalar
Yukawa theory, with Lagrangian
1
L = (∂φ)2 − m2 φ2 + (∂µ ψ ∗ )(∂ µ ψ) − M 2 ψ ∗ ψ − gψ ∗ ψφ . (5.12)
|2 {z } | {z } | {z }
interaction
Complex Scalar
Klein-Gordan
This is the interaction between a Klein-Gordan field of mass m and a complex scalar field of
mass M .
ψ
p1 (p1 + p2 )
φ
p2 g
Exercise
Find the equations of motion for Equations (5.11) and (5.12).
As the result of the previous exercise shows, our equations of motion are non-linear and
so we can no longer solve them by Fourier decomposition. This is exactly what we wanted,
and will lead to different modes coupling together. As we will see, we solve these theories in a
perturbative way; we assume the couplings λ/g are small and expand around the free theory.
We are saved from hideously long expression by the famous Feynman diagrams. These things
really are a field theorist’s best friend.5
5
Sorry Harmonic Oscillators, you’ve been pipped at the post.
6 | Interaction Picture & Dyson’s For-
mula
So far we have used the Schrödinger picture and the Heisenberg picture. Recall that the
difference between these two pictures was where the time dependence sat: for the Schrödinger
picture the states are time dependant, whereas in the Heisenberg picture it is the operators
that are time dependant. There is a third picture, the interaction picture, which is a sort of
hybrid of the other two. As the name suggests, it is useful when studying interacting systems.
What we do is consider small perturbations from some well-understood Hamiltonian, H0 .
That is, we write the Hamiltonian as
H = H0 + Hint .
We treat Hint like a Schrödinger Hamiltonian, and H0 like a Heisenberg one. That is, the
time dependence of the states is governed by Hint while the time dependence of the operators
is governed by H0 . As we did for the other pictures, we denote the states/operators in the
interaction picture with a subscript I, and they are given as follows:
|ψ(t)iI = eiH0 t |ψ(t)iS , and OI (t) = eiH0 t OS e−iH0 t . (6.1)
From this we ca work out how the states |ψiI depend on time. We simply use the Schrödinger
equation
d
i |ψiS = HS |ψis
dt
along with HS = (H0 + Hint )S . We get
d −iH0 t
|ψiI = H0 + Hint S e−iH0 t |ψiI
i e
dt
−iH0 t
−iH0 t d
i |ψiI = H0 + Hint S e−iH0 t |ψiI ,
H0 e |ψiI + e
dt
so we get
d
i |ψiI = (Hint )I |ψiI , with (Hint )I := eiHo t (Hint )S e−iH0 t . (6.2)
dt
Notation. From now on we shall simply write HI (t) instead of (Hint )I (t) for obvious reasons.
However we should be careful not to confuse HI (t) with Hint itself: the former is the latter in
the interaction picture, while the latter is a small perturbation to our Hamiltonian. Note in
particular that the former is always a function of time whereas the latter is time independent
in the Schrödinger picture.
45
LECTURE 6. INTERACTION PICTURE & DYSON’S FORMULA 46
Exercise
Convince yourself that [HI (t1 ), HI (t2 )] 6= 0 in general.
Ok this looks great but it’s useless unless we can actually solve it, so what do we do? The
answer is we derive what is known as Dyson’s formula.
d
i U (t, t0 ) = HI (t)U (t, t0 ). (6.3)
dt
Now if we were just considering wavefunctions here instead of operators then the solution to
this differential equation would just be
Z t
0 0
U (t, t0 ) = exp − i HI (t )dt . (6.4)
t0
The second term is the kind of thing we want but the first term has the HI (t) on the wrong
side! We cannot simply ‘pull it through’ as the operators are non-commuting. So we need to
alter Equation (6.4) somehow to account for this. The obvious2 thing to try is taking a time
ordering.
1
As with all good physics derivations do, our guess will turn out to miraculously be correct...
2
Or perhaps only obvious when you know its the answer...
LECTURE 6. INTERACTION PICTURE & DYSON’S FORMULA 47
Remark 6.1.2 . Before presenting the proof, let’s make a quick remark. Recall that if the result
of an integral is finite, then we can use3
Z b 2 Z b Z b
f (x)dx = dx dyf (x)f (y).
a a a
Now in QFT we consider bounded operators, and so the integral over the Hamiltonian is
finite, so we can turn our exponential expansion above into a sum of higher order integrals,
rather then powers of integrals. This is obviously a massive help. In particular, our expansion
becomes
Z t "Z Z t0 #
t Z t Z t
0 0 (−i)2 0 00 00 0 0 00 0 00
U (t, t0 ) = 1−i dt HI (t )+ dt dt HI (t )HI (t )+ dt dt HI (t )HI (t ) +...
t0 2 t0 t0 t0 t0
Note the integration limits on the quadratic terms. The first one is the condition that t00 > t0
as the lower limit for t00 is t0 , while the second one is the condition that t0 > t00 as the upper
limit for t00 is t0 . These two terms are in fact equal as
Z t Z t Z t Z t00
0 00 00 0 00
dt dt HI (t )HI (t ) = dt dt0 HI (t00 )HI (t0 )
t0 t0 t0 t0
Zt Z t0
= dt0 dt00 HI (t0 )HI (t00 ),
t0 t0
where the left had side says t00 has lower limit t0 while the first expression on the right hand
side says t0 has supper limit t00 . These are clearly the same statement. The last expression is
obtained by simple change of variables t0 ←→ t00 . We can also see this diagrammatically as
the two following triangle areas: the blue area corresponds to the left hand side above (i.e.
t00 > t0 ), whereas the red area is the final expression (i.e. t0 > t00 ).
t00
t0
t0 t t0
3
If you haven’t seen this before, it’s worth convincing yourself why this is true.
LECTURE 6. INTERACTION PICTURE & DYSON’S FORMULA 48
Proof. Given the final result from the above remark, the proof of Dyson’s formula is almost
trivial. We simply note that our left most integral is always the one with upper limit t, and
so when we act with the time derivative we are always just getting the latest time result (as
the expression is time ordered). We can therefore pull it out of the time ordering and get the
result. In terms of maths, that is
" Z t # Z t #
d 0 0 0 0
i T exp − i HI (t )dt = T HI (t) exp − i HI (t )dt
dt t0 t0
" Z t #
0 0
= HI (t) T exp − i HI (t )dt ,
t0
Dyson’s formula might look a bit daunting, but when we remember that in the interaction
picture we’re considering small perturbations, we can truncate our expansion and it becomes a
lot nicer. Now, recalling that we said in order to study interacting filed theories we were going
to take a perturbation around a free theory, we see why Dyson’s formula is so useful. It will
allow us to use something called Wick’s theorem and then define the all-so-useful Feynman
diagrams. But first let’s talk about scattering.
6.2 Scattering
Definition. We define the scattering matrix, or S-matrix for short, to be the amplitude to
go from some initial state |ii to a final state |f i:
The initial and final states, which we collectively call asymptotic states, are
eigenstates of the free Hamiltonian. .
The basic idea is we want to say that the initially the fields/particles are so far apart that
they do not interact with each other at all, and so do not feel in the interaction Hamiltonian.
As Prof. Tong points out,4 at first this seems like a reasonable thing to do, but then we
realise it’s actually not as solid an assumption as we might think. We shall illustrate one of
the problems here.
4
Pages 54-55.
LECTURE 6. INTERACTION PICTURE & DYSON’S FORMULA 49
Consider the case of a real Klein-Gordan field, φ, in some potential V (x), the Lagrangian
is
L = LKG − V (x)φ,
then the equations of motion become
(∂ 2 + m2 )φ = −V (x).
So our asymptotic states condition tells us that we need V (x) → 0 as |x| → ∞. This still
seems somewhat reasonable. However now let’s consider the interaction term to be a φ4 term.
Recall the Lagrangian is
λ
L = LKG − φ4 ,
4!
and you should have shown in the exercise that the equation of motion is
λ
(∂ 2 + m2 )φ = − φ3 .
3!
This poses a more significant problem, as we can’t ‘turn off’ our interaction at |t| → ∞;
the field is defined everywhere! Despite this we continue on with our assumption that the
asymptotic states are eigenstates of the free theory. We therefore rewrite Equation (6.7) as
where the states are now eigenstates for the free theory.
We can actually define a different matrix, called the transition matrix, by
Tf i := hf | (S − 1) |ii ,
which removes the 1 term in the Taylor expansion of U (t, t0 ).
We’re going to ‘flip’ the diagram we gave for this at the end of last lecture, and consider the
decay process
φ(p) → ψ(q2 )ψ(q1 ),
where φ is the particle created by a† , ψ the one created by b† and ψ the antiparticle created
by c† . The diagram6 looks like
5
This notation is meant to mean everything in the bracket is a function of x. Just want to save writing all
the brackets and clutterig notation.
6
We will understand how to draw these soon.
LECTURE 6. INTERACTION PICTURE & DYSON’S FORMULA 50
ψ
p
q1
φ
g q2
So what is our scattering amplitude? Well we are just considering the first order expansion,
i.e. one power of HI (higher powers would correspond to terms like ψ(p1 )ψ(p2 ) → φ(q) →
ψ(p3 )ψ(p4 ) etc), so we have
Z
U (−∞, ∞) = 1 − ig d4 xψx† ψx φx + O(g 2 ),
We now need to substitute in ψx† , ψx and φx . First let’s just do φx . We use the definition
d3~k
Z
1 −ikx † ikx
φ(x) = a~ e + a~k e ,
(2π)3 2E~k k
p
and note that the a~† term will act on |ii and produce a two particle state. That is (up to
√ k
factors of 2E)
a~† |ii = a~† ap†~ |0i = |φ(p)φ(k)i .
k k
Then noting that the ψ/ψ †terms are not going to be able to ‘undo’ this, we will get a zero
projection onto the final state |f i. So we just need to consider the a~k term. So using the
usual trick of commuting a~k with the ap†~ and picking up a (2π)3 δ (3) (~k − p~), we get
Z
Tf i = −ig hf | d4 x ψx† ψx e−ipx |0i .
Exercise
Hint: Again you will argue away terms that will vanish when taking the inner product.
This is why I have put “=”.
LECTURE 6. INTERACTION PICTURE & DYSON’S FORMULA 51
This result is basically just telling us that 4-momentum must be conserved at the vertex, a
very nice physical result. What kind of restrictions does this put on the process? Well the
result is Lorentz invariant so we can pick the decaying φs rest frame to evaluate it. In this
frame we have p = (m, 0, 0, 0), so the delta function splits into
δ (4) (p − q1 − q2 ) = δ(m − q10 − q20 )δ (3) (~q1 + ~q2 ),
and so we see the decay can only happen if m ≥ 2M .
Obviously we expect momentum conservation to be something that comes up a lot in our
interaction calculations, and indeed it turns out that in general
Tf i = i(2π)4 Mf i δ (4) (pi − pf ), (6.8)
where pi and pf are the total 4-momentum initial and final state, respectively. So essentially
our goal is to find the Mf i s, and these are given the, somewhat disappointing, name matrix
elements. We often also group the i with it and call iMf i the matrix elements.
6.2.2 φ4 Scattering
Ok so we’ve computed our first scattering matrix, but it corresponded to a decay, so now let’s
find the first scattering scattering matrix. That is we want something like
F (p1 )G(p2 ) → F (p3 )G(p4 ),
where F and G are some fields. We could consider higher order terms in the expansion for the
scalar Yakawa theory as mentioned above, however as we will see these are incredibly hard to
compute and will require the machinery of Wick contractions. So we want something that to
first order in HI will give us 4 fields (two in, two out). The easiest example is of course φ4
theory, which has Z
λ
Hint = d3 x φx φx φx φx ,
4!
and corresponds simply to the scattering process
φ(p1 )φ(p2 ) → φ(p3 )φ(p4 ).
The calculation of the transition matrix is left as an exercise.
Exercise
Show that the above interaction Hamiltonian leads to
Note that the factor of 4! has gone in the result to the above exercise. This is actually
the reason it’s included in the original expression, so that the matrix elements don’t have any
additional factors. Its an example of a what we call symmetry factors, which will become
more clear after studying Wick contractions.
LECTURE 6. INTERACTION PICTURE & DYSON’S FORMULA 52
Remark 6.2.1 . Note that for the scalar Yakawa decay we could have actually used the S-matrix
instead of the transition matrix. This is because the two only differ by the inner product
between the initial and final state, which for the scalar Yakawa theory vanishes. However for
the φ4 scattering, this inner product is non-vanishing and doesn’t actually correspond to a
scattering. Diagrammatically it looks like,
p1 = p3
φ φ
φ φ
p2 = p4
Last lecture we showed how we could use Dyson’s formula to find the S-matrix for a simple
decay and simple scattering. We made a comment about how would could have got a scattering
in the scalar Yakawa theory, but that it would involve a quadratic term from Dyson’s formula,
and so would be very hard to do in the manner above. However, obviously we want to be
able to calculate such terms (otherwise we never pass first order in perturbation theory, and
we loose a lot of information), so how do we do it? The answer is we be smart about how we
write things.
Ok we just said we "be smart", but what does that actually entail? We shall use the scalar
Klein-Gordan field to explain. Recall that we can write this field as the expansion
d3 p~
Z
1 −ipx † ipx
φ(x) = a p
~ e + ap~ e .
(2π)3 2Ep~
p
Remark 7.1.1 . It might seem strange that we define φ+ (x) to be the one that has the negative
exponential. The reason for this is that e−ipx corresponds to a mode with positive frequency,
and so we use φ+ (x). Also note that the dagger does not come with φ+ (x), as you might think.
This is actually important because it is fact that φ± (x) correspond to annihilation/creation
operators, respectively, that we will use.
Now, as we just mentioned in the remark above, φ+ (x) comes with an annihilation oper-
ator, so if it were to appear to the far right in a vacuum expectation value, the result would
be zero. The same thing is true for φ− (x) being to the far left. That is
53
LECTURE 7. WICK’S THEOREM 54
pi = e−ipx |0i ,
φ+ (x) |~ and p| φ− (x) = eipx h0| .
h~
We call the procedure of putting all the annihilation operators to the right normal ordering.
: AB + CD : = : AB : + : CD : .
Remark 7.1.3 . Note we could have introduced this (as Prof. Tong does) when talking about
the energy of the ground state being infinite. That is we could write the Hamiltonian for the
single free scalar field as
d3 p~
Z
: H: = Ep~ ap†~ ap~ ,
(2π)3
so that
: H : |0i = 0.
Notation. The colon notation used here for normal ordering is very common, however some
authors, including Peskin and Schroeder, denote normal ordering by an N . So Equation (7.2)
would be written N (O1 (x1 )...On (xn )).
and let’s consider x01 > x02 . Then the time ordered product is1
T φ1 φ2 := φ1 φ2
− + −
= (φ+
1 φ1 )(φ2 φ2 )
+ − − + − −
= φ+ +
1 φ2 + φ1 φ2 + φ1 φ2 + φ1 φ2
− + + − − + − −
= φ+ +
1 φ2 + φ2 φ1 + [φ1 , φ2 ] + φ1 φ2 + φ1 φ2
−
= : φ1 φ2 : + [φ+
1 , φ2 ]
= : φ1 φ2 : + D(x1 − x2 ),
where we have used the commutator to make every term have a φ+ to the far right and the
definition of the propagator D(x − y).
Exercise
Following a similar method to the above, show that if x02 > x01 , then
T (φ1 φ2 ) = : φ1 φ2 : + D(x2 − x1 ).
Hang on, this looks familiar... the cases term is just the Feynman propagator, Equation (5.8).
So we get
T (φ1 φ2 ) = : φ1 φ2 : + ∆F (x1 − x2 ). (7.3)
Remark 7.2.1 . Note that once again we have an expression that mixes operators and C-
numbers: both T (φ1 φ2 ) and : φ1 φ2 : are operators, but the Feynman propagator is a C-
number. We have made this comment several times throughout this course, but make it here
again to keep us on our toes.
1
Be careful not to confuse the colon in :=, which means "is defined to be equal to" to the colons for normal
ordering. I considered using another type font for the colon, but decided against it.
LECTURE 7. WICK’S THEOREM 56
where the ... on the right hand side does not include φi and φj .a
a
If you like, in standard "this element is missing" notation, we could write the right hand side as
∆F (xi − xj )φ1 ...φ̂i ..., φˆj ...φn , but I thought this would be confusing, as hats are used for operators
in QFT.
T (φ1 φ2 ) = : φ1 φ2 : + φ1 φ2 .
Claim 7.2.3 . A similar argument/proof can be made for complex fields, where we have
so that
T (ψ1 ψ2† ) = : ψa ψ2† : + ψ1 ψ2† .
So we see that for the case of two real scalar fields, and claimed it for two complex
scalar field, the difference between the time ordered and normal ordered products is a Wick
contraction, but what about if we have more than two fields? Well this is where Wick’s
theorem comes in.
Theorem 7.2.4 (Wick’s Theorem). For any collection of fields φ1 , ..., φn the following
holds
T (φ1 ...φn ) = : φ1 ...φn : + : all possible contractions : . (7.5)
Proof. We do not prove Wick’s theorem here, but an explanation of the proof (its an inductive
proof) is given in Prof. Tong’s notes, page 58.
It is often very useful to break a calculation using Wick’s theorem into categories, given by
the number of contractions done. Every calculation will have the 0-contraction term, which
is just the fully normal ordered term, and then you consider the terms in increasing number
of contractions.2
Example 7.2.5 . Let’s consider 4 real Klein-Gordan fields. Then we can have either 0-contractions,
1-contraction or 2-contractions. The respective terms are:
• 0-contractions
: φ1 φ2 φ3 φ4 : ,
2
Well it depends what you want to do. In string theory it is often the term which is fully contracted that
contains most of the useful information. For more details, see my notes on Dr. Shiraz Minwalla’s string theory
course.
LECTURE 7. WICK’S THEOREM 57
• 1-contraction:
: φ1 φ2 φ3 φ4 + φ1 φ2 φ3 φ4 + φ1 φ2 φ3 φ4 + φ1 φ2 φ3 φ4 + φ1 φ2 φ3 φ4 + φ1 φ2 φ3 φ4 :
• 2-contractions
: φ1 φ2 φ3 φ4 + φ1 φ2 φ3 φ4 + φ1 φ2 φ3 φ4 :
h0| T (φ1 φ2 φ3 φ4 ) |0i = ∆F (x1 −x2 )∆F (x3 −x4 )+∆F (x1 −x3 )∆F (x2 −x4 )+∆F (x1 −x4 )∆F (x2 −x3 ),
which is a Green’s function for the Klein-Gordan operator (recall that the Feynman integral
is a Green’s function).
pi = e−ipx ,
φ(x) |~ and p| φ(x) = eipx ,
h~ (7.6)
and then Wick’s theorem is simply adapted to include contractions between fields and states.
This allows us to calculate the S-matrix completely using Wick’s theorem.
We can now return to the problem of trying to find the O(g 2 ) term in scalar Yakawa theory.
Recall the Lagrangian is,3
1 1 1
L = (∂φ)2 − m2 φ2 + ∂µ ψ † ∂ µ ψ − M 2 ψ † ψ − gφψ † ψ.
2 2 2
Similarly to Equation (7.1) we define
d3 p~ d3 p~
Z Z
1 1
+
φb (x) = bp~ e −ipx
, and φ −
b (x) = b† eipx
3 3 2Ep~ p~
p p
(2π) 2Ep~ (2π)
d3 p~ d3 p~
Z Z
1 1
+
φc (x) = c p
~ e −ipx
and φ −
c (x) = c† eipx
3 3 2Ep~ p~
p p
(2π) 2Ep~ (2π)
so that
− −
ψ(x) = φ+
b (x) + φc (x), and ψ † (x) = φ+
c (x) + φb (x),
where we note that each term contains a b and a c, in correspondance with Equation (5.1).
Then, we recalling the definitions
3
Using ψ † now instead of ψ ∗ because we will only consider the quantum theory from now on.
LECTURE 7. WICK’S THEOREM 58
ψ(x) |ψ(p)i = ψ † (x) ψ(p) = e−ipx , and hψ(p)| ψ † (x) = ψ(p) ψ(x) = eipx ,
(−ig)2
Z
† †
d x d y ψx ψx φx ψy ψy φy |ψ(p1 )ψ(p2 )i + O(g 3 ).
4 4
+ ψ(p3 )ψ(p4 ) T
2!
We can use Wick’s theorem to evaluate this. First we note that the first term vanishes. Why?
Well because we have nothing to contract the φx with, so we have to leave it as a normal
ordered term, which then acts on the vacuum giving a vanishing result. We will see in terms
of Feynman diagrams soon why such a term can’t exist. Now what about the second term?
Well we get three different types of contractions which are conceptually different they are
(iii) Disconnected.
We shall consider the fully connected diagrams in this lecture and derive the corresponding
Feynman rules, and then talk about connected and disconnected diagrams next lecture.4
g2
Z
d4 x d4 y ψ(p3 )ψ(p4 ) : ψx† ψy† φx φy ψx ψy : |ψ(p1 )ψ(p2 )i + (x ←→ y),
−
2!
and
g2
Z
d4 x d4 y ψ(p3 )ψ(p4 ) : ψx† ψy† φx φy ψx ψy : |ψ(p1 )ψ(p2 )i + (x ←→ y).
−
2!
The (x ←→ y) terms are symmetries factors that will remove the 1/2 factors, just like ones we
mentioned before that got rid of the 1/4! in φ4 theory. We will see soon in terms of Feynman
diagrams why they are symmetry factors.
Evaluating the contractions using the rules we have above, these two terms become, re-
spectively Z h i
d4 x d4 y ei(p3 −p1 )x (−ig)∆φF (x − y)(−ig) ei(p4 −p2 )y , (7.7)
4
Just because that’s what happened on the course.
LECTURE 7. WICK’S THEOREM 59
and Z h i
d4 x d4 y ei(p4 −p1 )x (−ig)∆φF (x − y)(−ig) ei(p3 −p2 )y . (7.8)
Exercise
Check that the Wick contractions do indeed result in the above two expressions.
The position space Feynman rules for scalar Yakawa scattering ψψ → φ → ψψ are:
Propagator x y
∆φF (x − y) = φx φy
x
Vertex d4 x
R
ig
• We take time to run from left to right. So the initial state is to the far left and final state
is to the far right. Note, though, that some authors choose to have time run vertically.
Peskin and Schroeder use this latter convention.
• We use solid lines to indicate complex fields (or nucleons) and dotted lines to indicate
real fields (or mesons).
• We use a filled in circle to indicate a vertex, and we label it by its integration variable.
• The arrow above the lines indicate the value of the 4-momentum and the direction it
flows.
LECTURE 7. WICK’S THEOREM 60
• The arrows on the nucleon line tells us about the flow of the charge (e.g. Fermion
number), if the line points in the same direction as the momentum arrow we have a
particle, and if they point oppositely, then we have an antiparticle. This is why we
don’t draw an arrow on the dotted line; a real scalar field is its own antiparticle.
• The arrows for a charge must always be conserved at a vertex. By this we mean one
must always go in and one must always leave.
• The incoming and outgoing particles are known as external line because they only
connect to one vertex, whereas the propagator is called an internal line because it is
connected to two vertices.
• We usually denote the 4-momentum on the external lines with pi , and in the position
space picture we don’t need to label the momentum of the internal lines.
• Sometimes we also include the symbols of the particles next to the lines. This is partic-
ularly the case when we have multiple different species particle of the same type (e.g.
electrons and muons) in the same diagram.
Using these rules, we can construct the Feynman diagrams for Equations (7.7) and (7.8).
They are, respectively:
p1 p3 p1
x x
ψ ψ ψ p3 ψ
φ φ
p4
ψ ψ ψ ψ
y y
p2 p4 p2
Remark 7.3.1 . The symmetry factor x ←→ y in terms of the Feynman diagrams is seen by
the fact that if we relabel x ←→ y then the diagram looks exactly the same apart from the
relabelling. So there is a symmetry in the diagram, hence the name.
Remark 7.3.2 . Note that we can see from the Feynman diagrams that the order g term wont
contribute, as we said before. That’s just because there is no way to start with two ψ lines and
end up with two ψ lines using only one vertex.5 This illustrates one of the incredible powers
of Feynman diagrams: it allows us to see almost instantly which terms do not contribute to
the transition matrix.
Exercise
Convince yourself that these diagrams are correct. Hint: This exercise should make it
clear why we collected the terms in Equations (7.7) and (7.8) ad we did.
5
If you don’t see this, try and do it and you’ll quickly see what I mean.
LECTURE 7. WICK’S THEOREM 61
d4 p ie−ip(x−y)
Z
∆F (x − y) = ,
(2π)4 p2 − m2 + i
and so we can replace the position integrals with momentum ones. This gives us the momen-
tum space Feynman rules. Let’s do this explicitly for Equation (7.7). Plugging the Feynman
propagator in, and using k as the momentum variable in correspondance with the rules above,
we have
d4 k ie−ik(x−y)
Z h Z i
4 4 i(p3 −p1 )x i(p4 −p2 )y
d xd y e (−ig) (−ig) e ,
(2π)4 k 2 − m2 + i
combining the x and y exponentials and then integrating over x and y gives us two delta
functions, that is we get
d4 k
Z
i
(2π)4 δ (4) p1 − p3 − k (−ig) 2 (−ig)(2π)4 δ (4) p2 − p4 + k ,
(2π) 4 2
k − m + i
i
iM = (−ig) (−ig).
(p1 − p3 )2− m2 + i
Exercise
Show that Equation (7.8) leads to
4 (4)
i
(2π) δ p1 + p2 − p3 − p4 (−ig) (−ig) ,
(p1 − p4 )2 − m2 + i
but note the difference of the intermediate steps. Hint: Be careful about delta functions.
So how do the Feynman rules change? Well we note now that x and y don’t appear in
the final expression so we don’t need to label the nodes any more. Then we note in the
second step, the two delta functions simply correspond to conservation of 4-momentum at the
vertices, i.e. δ (4) (p1 − p3 − k) is the momentum flowing into the x vertex above. Then to get
the final answer to integrated over the internal momentum k, which gave use the final delta
function.
6
Note we could also use (p2 − p4 )2 in the denominator.
LECTURE 7. WICK’S THEOREM 62
The momentum space Feynman rules for scalar Yakawa scattering ψψ → φ → ψψ are:
d4 k
Z
Propagator k i
(2π) k − m2 + i
4 2
−ig
Vertex −(ig)
So we have the rough7 following procedure to obtain the maths expression from a momentum-
space Feynman diagram:
• Take the position space Feynman diagram, and given every internal line a momentum
labelled by ki .
• Next to each vertex write down the coupling strength (in the case of scalar Yakawa this
is −ig).
• Impose momentum conservation at each vertex with a factor of the coupling strength,
i.e. write (−ig)(2π)4 δ (4) ( i ki ), where the sum is taken over the momentum flowing
P
into the vertex. That is, if the momentum flows out, put a minus sign in the sum, as
we did in δ (4) (p1 − p3 − k) and δ (4) (p2 − p4 + k).
• Include a factor of
d4 k
Z
i
(2π)4 ki2 − m2 + i
for all the internal φ lines, and the same for internal ψ lines but now with m → M .
R d4 k
Remark 7.3.3 . Some authors do not include the integral factor (2π) 4 in the propagator
term expression, but instead add another step at the end that says "finally integrate over all
internal momentum". Of course this amounts to exactly the same thing, but this remark is
just included to avoid confusion when comparing to other literature.
Exercise
Draw the scalar Yakawa scattering Feynman diagrams above in momentum space.
7
We will need to be more clear for the case of spin-1/2 particles later.
8 | LSZ Theorem, Loop & Amputated
Diagrams
Last lecture we discussed fully connected diagrams and derived the Feynman rules for ψψ →
ψψ scattering. We also said that there were two other types of conceptually different diagrams,
namely connected and disconnected diagrams. As diagrams they look like:
ψ ψ ψ ψ
ψ ψ ψ
φ
ψ ψ ψ ψ
Connected Disconnected
Intuitively, we would say these don’t really correspond to scattering as the initial states
don’t interact with each other. However we need some way to actually prove this, and this is
where the so-called LSZ theorem1 comes in. First we need some definitions and a proposition.
Definition (n-Point Green’s Function). We define the n-point Green’s function to be the
time ordered vacuum expectation value of n Heisenberg picture2 field operator. That is
(8.1)
Gn (x1 , ..., xn ) := h0| T φ(x1 )...φ(xn ) |0i .
hp1 , ..., pn | φ(t = ±∞, ~x) |q1 , ..., qm i = lim Z −1/2 hp1 , ..., pn | φ(x) |q1 , ..., qm i . (8.2)
t→±∞
1
Named after Harry Lehmann, Kurt Symanzik and Wolfhart Zimmermann.
2
Note we need this, otherwise time-ordering doesn’t mean anything.
63
LECTURE 8. LSZ THEOREM, LOOP & AMPUTATED DIAGRAMS 64
Definition (Left-Right Derivative Action). 3 Let φ1 (x) and φ2 (x) be to fields on the space-
time. Then we define the left-right derivative action as
↔
(8.3)
φ1 (x) ∂ µ φ2 (x) := φ1 (x)∂µ φ2 (x) − ∂µ φ1 (x) φ2 (x).
Proposition 8.0.2. For the real scalar field, Equation (4.10), we have
↔
Z
i
ap~ = p d3 ~x eip·x ∂ 0 φ(x)
2Ep~
Z ↔
(8.4)
† i
ap~ = − p d3 ~x e−ip·x ∂ 0 φ(x)
2Ep~
Remark 8.0.3 . Note that because we only integrate over space but φ(x) is a function of space-
time, our creation and annihilation operators are functions of time! This will prove crucial in
the proof of the LSZ theorem below.
Exercise
Prove Equation (8.4). Hint: Youpwill want to use the constraint p0 = Ep~ present in
Equation (4.10) along with Ep~ = p~2 + m2 = E−~p at the end.
We choose here to first state the result of the LSZ theorem and then derive it, as this way
the heavy manipulations that follow have some guiding light.
Theorem 8.0.4 (LSZ Reduction Formula). We can express the S-matrix4 between an initial
m-particle state and a n-particle final particle state in terms of the (n + m)-point Green’s
function as follows:
Z Z
−1/2 n+m
dy1 ...dyn dx1 ...dxm ei(pj ·yj −qi ·xi )
Tf i = iZ
Ym Y n
× 2 2
(∂xi + m ) 2
(∂yj + m ) Gn+m (y1 , ..., yn , x1 , ..., xm ) (8.5)
i=1 j=1
Before proving this theorem it is worth stating what it allows us to do. Basically we
note that the right-hand side of Equation (8.5) splits into a contribution that contains only
information about the fully connected diagrams (we will prove this), and a part that contains
only the non-fully connected stuff. We can therefore simply restrict our analysis to processes
3
This might not be a technical name, it’s just what I’ve decided to call it.
4
Again we are really interested in the Transition matrix. I have tried to use Ts here, but I might have
missed some, and written S by accident. Apologies for that, please feel free to point these out to me.
LECTURE 8. LSZ THEOREM, LOOP & AMPUTATED DIAGRAMS 65
that are only fully connected in a well defined way. That is, a priori there is no clear way to
see that the result of the full S-matrix calculation wont ‘mix’ fully connected and non-fully
connected stuff, but the LSZ theorem tells us that it indeed does. This statement does not
mean that the connected diagrams do not contribute to the full S-matrix value, but simply
that we are allowed to restrict ourselves to considering only fully-connected diagrams.
Proof. Unfortunately for those who don’t like long proofs, this one comes from sheer brute
force,5 so prepare yourself. Firstly we need to clarify the notation that follows: by Dysons for-
mula, the S-matrix is given by a bunch of time evolution operators, and we also take our states
to asymptotic (i.e. at t → ±∞). Therefore the creation operators that we extract from our
initial states do not simply act on the final state and annihilate them, as they are separated
in time. For this reason we shall use subscripts ‘in’ and ‘out’ on the creation/annihilation op-
erators to avoid confusion. Similarly we will denote the initial/final states using the following
notation
hp1 ...pn ; out| , and |q1 ...qm ; ini .
We use this notation as it is the one used by Prof. Weigand in his notes.
We therefore have
Tf i = hp1 ...pn ; out|q1 ...qm ; ini
= 2Eq~1 hp1 ...pn ; out| a†q~1 |q2 ...qm ; ini
p
↔
Z
1
= d3 ~x1 hp1 ...pn ; out| e−iq1 ·x1 ∂ 0 φ(t = −∞, ~x) |q2 ...qm ; ini
i
↔
Z
1 −1/2
= lim Z d3 ~x1 hp1 ...pn ; out| e−iq1 ·x1 ∂ 0 φ(x) |q2 ...qm ; ini
i t→−∞
↔
Z
1
= √ lim d3 ~x1 e−iq1 ·x1 ∂ 0 hp1 ...pn ; out| φ(x) |q2 ...qm ; ini
i Z t→−∞
where we have used the definition |qi = 2Eq~a†q~ |0i and Equations (8.2) and (8.4). Note on
p
the third line we have φ(t = −∞, ~x), this is because we want to consider an initial state,
which we take to be at t = −∞. Indeed for this exact reason, we should really denote the
creation operator on the second line as something like
(ain )†q~1 ,
so that we know it only acts on in/initial states.
What we now want to do is get the φ(x) to act on a final state, in which case we need to
change the limit from t → −∞ to t → +∞. We achieve this by using the simple result
Z T
d
lim f (t) = lim f (t) − lim f (t),
t→−∞ t→+∞ T →∞ −T dt
where we notice that the second term contains a 4-integral. Let’s call the first A and the
second B, and consider them in turn. In A we can use Equation (8.2) backwards to get
↔
Z
1
A= √ d3 ~x1 e−iq1 ·x1 ∂ 0 hp1 ...pn ; out| φ(t = +∞, ~x) |q2 ...qm ; ini ,
i Z
and then we can use Equation (8.4) backwards to give
where, in agreement with the comment made above, we have included a subscript ‘out’ to tell
us this acts on the final states. Now the action of a creation operator to the left gives a delta
function, and so we get
n
X
3
A = 2Eq~1 (2π) δ (3) (~
pk − ~q1 ) hp1 ...p̂k ...pn ; out|q2 ...qm ; ini ,
k=1
where we have used the standard maths notation where a hatted entry in a string of terms
is missing (i.e. p̂k is missing from the final states). This corresponds to a connected, but not
fully connected, diagram as we have the k-th final particle just corresponding to the 1st initial
particle (it is just a straight line in a diagram).
Now what about B? Well we use Equation (8.3) to get
Z
1
d4 x1 ∂0 e−iq1 ·x1 ∂0 h...i − ∂0 e−iq1 ·x1 h...i
B= √
i Z Z
1
d4 x1 e−iq1 ·x1 ∂02 h...i − ∂02 e−iq1 ·x1 h...i
= √
i Z
Next note that
which we can apply to the second term on the second line above. We then integrate by parts
twice6 to move the ∇2 to act on h...i and then use
∂ 2 := ∂02 − ∇2
to give us Z
1
B= √ d4 x1 e−iq1 ·x1 ∂ 2 + m2 ) hp1 ...pn ; out| φ(x) |q2 ...qm ; ini
i Z
This is exactly of the form of one of the product terms in Equation (8.5)! We can then apply
a similar argument to all the other initial states, noting that
We now need to do the same thing for the final states. However now we have
p
hp| = 2Ep~ h0| (aout )p~ ,
and from
[ap~i , a†q~j ] 6= 0
we cannot simply commute the annihilation operators past φ(x) to get them to act on the
initial state, that is
hp2 ...pn ; out| φ(y10 = ∞,~y1 )φ(x01 = −∞, ~x1 ) |q1 ...qm ; ini
6= hp2 ...pn ; out| φ(x01 = −∞, ~x1 )φ(y10 = ∞, ~y1 ) |q1 ...qm ; ini .
This is when we notice that Equation (8.5) contains a Green’s function, which contains time-
ordering and so we realise we are saved! More explicitly, we will get something of the form
↔ ↔
Z Z
lim lim d3 ~x1 d3 ~y1 e−iq1 ·x1 ∂ x01 eip1 ·y1 ∂ y10 hp2 ...pn ; out| φ(y1 )φ(x1 ) |q2 ...qm ; ini
x01 →−∞ y10 →+∞
↔ ↔
Z Z
−iq1 ·x1
= lim 3
d ~x1 e ∂ x01 lim d3 ~y1 eip1 ·y1 ∂ y10 hp2 ...pn ; out| φ(y1 )φ(x1 ) |q2 ...qm ; ini
x01 →−∞ y10 →+∞
where we notice that the x01 limit is taken after the y10 one, so during the latter we can take
x01 to be some fixed finite value, so we can just consider the y1 part of the expression. Now
consider
Z T Z
↔
0 3 ip1 ·y1
lim dy1 ∂y10 d ~y1 e ∂ y10 hp2 ...pn ; out| T φ(y1 )φ(x1 ) |q2 ...qm ; ini
T →∞ −T
↔
Z
= lim d3 ~y1 eip1 ·y1 ∂ y10 hp2 ...pn ; out| φ(y1 )φ(x1 ) |q2 ...qm ; ini
y10 →+∞
↔
Z
− lim d3 ~y1 eip1 ·y1 ∂ y10 hp2 ...pn ; out| φ(x1 )φ(y1 ) |q2 ...qm ; ini .
y10 →−∞
The first term on the right hand side is what we want, (compare it to the proof above), while
the second term on the right hand side will again give a connected, but not fully connected,
diagram as we get a ap~1 term acting on the initial state. So we finally see that we can replace
the expression we had above for the Green’s function (the time-ordered expression) at the cost
of picking up a connected, but not fully connected, term. Doing this for all the final states will
then give us all the terms in the product from j = 1 to j = n needed in Equation (8.5). Note
also that all the other factors needed for Equation (8.5) also appear, i.e. all the exponentials
and the iZ −1/2 factors.
Ok so we have that the fully connected part of the scattering process is given by a (n+m)-
point Green’s function. However the LSZ formula is quite a large formula and so we want to
simplify it somehow. We do this by introducing the truncated Green’s function as follows
Z
Gn+m (y1 , ..., yn , x1 , ..., xm ) = d4 z1 ...d4 dzm+n ∆F (x1 − z1 )...∆F (xm − zm )
(8.6)
× ∆(y1 − zm+1 )...∆F (yn − zn+m )G e n+m (z1 , ..., zn+m ),
LECTURE 8. LSZ THEOREM, LOOP & AMPUTATED DIAGRAMS 68
where ∆F is the Feynman propagator and G e n+m is our truncated (n + m)-point Green’s
functions. Essentially what we’re doing is saying is propagate our initial states to the points
{z1 , ..., zm } and similarly the final states with {zm+1 , ..., zm+n }, and characterise all the in-
teraction stuff in the middle using some Green’s function, G, e as we’ve tried to indicate below
diagrammatically.
y1
x1
x2 y2
z6
z1
z2 z7
x3 z3 G
e5+4 z8 y3
z4 z9
z5
x4
y4
x5
Why would we introduce the truncated Green’s function? Well we recall that the Feynman
propagator is a Green’s function of the Klein-Gordan operator:
(∂ 2 + m2 )∆F (x − z) = 0,
so in the LSZ formula all of the Klein-Gordan operators will just give us delta functions,
which we can integrate over, leaving us with
Z Z
−1/2 n+m
TfF.C. dx1 ...dxm ei(pj ·yj −qi ·xi ) G
i = iZ dy1 ...dy n
e n+m (x1 ..., xm , y1 , ..., yn )
(8.7)
n+m
= iZ −1/2 G
e n+m (q1 , ..., qm , p1 , ..., pn ),
where the second line follows from the delta functions we get from the exponentials, and where
the superscript F.C. means "fully connected". This result tells us that we can compute the
entire fully connected S-matrix by just considering the momentum space truncated Green’s
function! This is a massively convenient result.
Remark 8.0.5 . It is worth stressing again that the LSZ theorem does not tell us that the
connected, but not fully connected, diagrams vanish but that we can ignore them if we want
to just consider a scattering process where everything interacts. In fact the former diagrams
will contribute to renormalisation and so are very important.
So far we have only considered what are known as tree level diagrams. This essentially just
means that all of our vertices are connected to our external lines. However this is clearly not
the only type of diagram we can have and we can have vertices internally. For example in the
scalar Yakawa theory the following is a valid diagram to order g 2 :
LECTURE 8. LSZ THEOREM, LOOP & AMPUTATED DIAGRAMS 69
p1 p3
ψ ψ
k1
`1
`2
k2
ψ ψ
p2 p4
where we haven’t labelled the internal particles and the (−ig) factors to avoid cluttering the
diagram. This is called a loop diagram for obvious reasons. Loop diagrams are a bit of a pain
because we don’t have enough delta functions to remove all of our integrals. That is, recall
that in the momentum space Feynman rules we include a factor of7
d4 ki
Z
i
(2π)2 ki2 − m2φ/ψ + i
for each internal line and a delta function of the momentum flowing in at each vertex. For
example, in the above diagram the k1 and k2 momentum are obtained from the delta functions
with the external momenta, i.e. k1 = p1 − p3 and k2 = p4 − p2 . However we cannot find the
values of `1 and `2 , but at best can obtain a relation between them. This is just because `1
and `2 will both appear in two delta functions, namely
δ (4) (k1 + `1 − `2 ), and δ (4) (`2 − k2 − `1 ),
and so when we integrate over one of the two `s we kill both delta functions and we are left
with the integral over the other ` and no delta function to remove it! To be more explicit, if
we did the integral over `2 , say, we would get
`2 = k1 + `1 , and `2 = k2 + `1 ,
so if we equate the two terms we just get
k1 = k2 =⇒ p1 + p2 = p3 + p4 ,
where the implication comes from the momentum conservation on the ks. This just tells us
the 4-momentum initially and finally are equivalent, and does not allow us to conclude what
`1 is.
Remark 8.1.1 . As we said when we introduced the momentum space Feynman rules, it is
common to alter the rules slightly to so just include the factor
i
k2 − m2 + i
for each propagator and then impose momentum conservation at each vertex. If we use these
rules, then when we have loops we insert the additional rule of "and then integrate over all
undetermined momenta", i.e. `1 in the above example.
7
Note that before we only defined this for φ propagators, but here we say we get the same factor for ψ ones
but we simply replace the mass with the ψ mass. From now on we shall use mφ = m and mψ = M .
LECTURE 8. LSZ THEOREM, LOOP & AMPUTATED DIAGRAMS 70
We also get loop diagrams in connected, but not fully connected, diagrams, as in the next
exercise.
Exercise
Show that this diagram
ψ ψ
p1 ` p2
corresponds to
"Z #
d4 `
i i
(2π)4 δ (4) (p2 − p1 ) (−ig)2
(2π)4 (p1 − `)2 − m2 + i `2 − M 2 + i
Remark 8.1.2 . Note that the result of the previous exercise diverges logarithmically in `. This
is easily seen because we essentially have 1/`4 and we do d4 `. This kind of divergence is known
as a UV divergence.
Note in all the Feynman diagrams we have just drawn, we always draw the ingoing and
outgoing particles as ‘pure’ lines, by which I mean nothing is attached to them. An example
of something we haven’t drawn is the following:
The reason we have never considered these is because what we’ve been drawing are so-
called amputated diagrams. You get amputated diagrams by saying "can I cut through the
diagram and remove something by only cutting through one line?" If the answer is yes, do it.
So for the above example we would cut along the dotted red line below:
LECTURE 8. LSZ THEOREM, LOOP & AMPUTATED DIAGRAMS 71
which gives us the scattering diagram we drew before. It is important that you only cut
through 1 line, i.e. we can not amputate the following
The reason we amputate diagrams is because the LSZ theorem tells us to consider only
amputated diagrams. This is seen from the truncated Greens function expression: we consider
just propagating our initial and final states in/out without any interactions and then only
consider the interactions that happen between all the particles in the internal parts, see the
truncated Ge picture above.
For completeness/have it all in one place, the table below gives the full set of (momentum
space) Feynman rules for Scalar Yakawa theory, the only real change to the table given before
is that we can now have antiparticle states ψ and we can have internal ψ/ψ states and external
φ states.
Here are some notes to go along with the table:
• As mentioned above, if you have a loop in the diagram, you will not have enough delta
functions to remove all the integrals, and so you will be left with an integral over the
loop momenta.
• Note that for the antiparticle ψ the charge flow and the momentum flow point in opposite
directions. This is how we distinguish particle from antiparticle in Feynman diagrams.
• Note that for the internal ψ propagator we don’t label the flow of charge (i.e. the arrow
on the line itself). This is because there is no way for us to test to see if it is a ‘particle
going forward in time’8 or an ‘antiparticle going backwards in time’. Obviously this is
just lingo to mean there is no way we can test to see which way the arrow points. In
either case we get the same result, and so we can just leave it as is. There is no harm
in putting the arrow, of course, just if you do you then have to make sure all the other
arrows internally agree, i.e. you must always have one arrow in and one arrow out at
every vertex.
8
Recall time flows left to right in our diagrams.
LECTURE 8. LSZ THEOREM, LOOP & AMPUTATED DIAGRAMS 72
• As mentioned at the end of the last bullet point, we always need one charge arrow in
and one charge arrow out of every vertex. In the ψψ → ψψ process this meant we could
not annihilated the two initial ψs (i.e. they didn’t meet at a vertex) as we would then
have two charge arrows into the vertex and zero out. However the ψs charge arrow
points oppositely to ψs, and so now they can annihilate. We draw this diagramatically
at the bottom of the table.
p
ψ
Outgoing External p
1
Lines ψ
p
φ
d4 k
Z
φ Propagator k i
(2π)4 k 2 − m2 + i
d4 k
Z
ψ Propagator k i
(2π)4 k 2 − M 2 + i
−ig
Vertices −(ig)
−ig
9 | φ4 Theory & Cross Sections
9.1 φ4 Theory
So we have looked at scalar Yakawa theory in some detail and derived the full set of Feynman
rules for this theory. We now want to do the other spin-0 theory we have been studying during
this course, φ4 theory. Recall that here we have
Z
λ
HI = d3 ~x φ4 (x),
4!
so we can use Dyson’s formula and Wick’s theorem to find the the S-matrix, or more correctly
T-matrix, as mentioned before. Let’s consider the scattering
Remark 9.1.1 . With the LSZ theorem in mind, we shall only consider fully connected di-
agrams, which corresponds to only considering terms where everything is contracted. This
should be reasonably clear, but if it is not, go back to the definition of what Wick contractions
give us and convince yourself this is true.
73
LECTURE 9. φ4 THEORY & CROSS SECTIONS 74
Exercise
Convince yourself that there are exactly 4! different permutations for the above con-
traction. Remark: As mentioned before, this result is exactly why we include the factor
of 1/4! in the Lagrangian in the first place, i.e. it is a symmetry factor.
Exercise
Show that the Wick contractions above lead to
iM = −iλ.
Hint: Use the result from the previous exercise along with Equation (7.6).
The Feynman rules rules are the same as before but now we have the vertex factor −λ.
The above scatting corresponds to the diagram
p1 p3
p2 −λ p4
We also get the terms with x ←→ y. Recall that the contraction between a φx and a φy
means that we have a φ propagating between x and y, so in position space the diagram looks
like2
y x
How do we find the number of permutations for this diagram? Well let’s just consider
the diagram as above, i.e. don’t consider switching x ←→ y, and then at the end we can
just double the answer at the end. So how do we find this? The answer is to go one Wick
contraction at a time. Let’s first considering the contractions with the final states: these
contract with φx s, of which there are 4, so the first contraction (say with p~4 ) has 4 choices,
2
We should label the momentum on the external states, but to save me doing some extra Tikz I’ve left
them out.
LECTURE 9. φ4 THEORY & CROSS SECTIONS 75
and then the second contraction (~p3 in this case, has 3 left options. To be a bit more explicit,
we have for the first contraction
h~
p4 p~3 | φx φx φx φx + h~
p4 p~3 | φx φx φx φx + h~
p4 p~3 | φx φx φx φx + h~
p4 p~3 | φx φx φx φx ,
and then for the second contraction each term above has three options, e.g. for the last term
we have
h~
p4 p~3 | φx φx φx φx + h~
p4 p~3 | φx φx φx φx + h~
p4 p~3 | φx φx φx φx .
So from this part we get 4 · 3 perturbations. Similarly the contractions of the initial states
with the φy s give another 4 · 3 perturbations. Then we have just left with the contractions
between the two remaining φx s and φy s, of which there are only two choices3
φx φx φy φy + φx φx φy φy .
So in total we have
(4 · 3) · (4 · 3) · 2 · 2 = 4!4!,
where the second 2 factor is the x ←→ y factor. So the numerical prefactor of the Wick
contraction is
1 4!4! 1
= ,
2 4!4! 2
so we get a symmetry factor of 1/2. Symmetry factors arise when elements of the diagrams
can be indistinguishably interchanged.
So far we have found the amplitude for an interaction to occur, but in quantum theory the
amplitudes themselves aren’t measurable. What we measure is the probability for a process to
occur, and that’s what we want to talk about now. However there appears to be a problem:
the probability is given by taking the absolute value squared of the amplitudes,
| hf | S − 1 |ii |2
P = , (9.1)
hf |f i hi|ii
delta function and so we consider the process happening in some finite time T , and then again
take the limit T → ∞. This will give us some formula for the probability, at which point
we have to realise that we cannot measure the final state momentum perfectly, but that they
will be smeared over some region in momentum space. To account for this, we must therefore
integrate over this region. We call this part of the formula the Lorentz invariant phase space
measure, or LIPS for short, and it is given by
Y d3 p~i 1
dΠ = (2π)4 δ (4) (pF − pI ) . (9.2)
(2π)3 2Ep~i
final states
We get the result for the differential probability per unit time to be
dP Y 1
= |iMf i |2 V dΠ (9.3)
T 2V Ep~i
initial states
where we note that the volume factors in Equation (9.3) have cancelled out. This is very
good because we want to take the limit V → ∞ without hitting any problems. We call dΓ the
differential decay rate probability.6 If we go to the particle’s rest frame we can replace E = m
(where m is the mass of our particle), giving
1
dΓ = |iMf i |2 dΠ
2m
To get the full decay probability, we need to consider the fact that our particle could decay
into a multitude of different states with different numbers (and/or types) or particles. This
gives us
Z
1 X
Γ= |iMf i |2 dΠ, (9.4)
2m
final states
known as the decay width of the particle. The decay width tells us the probability for the
particle to decay in any way shape or form, and so it is clearly related to the lifetime of the
particle by
1
τ= (9.5)
Γ
5
Note to self: maybe come back later and do the proof. Just don’t think its worth the time typing it out
now.
6
Or maybe differential decay probability per unit time? Basically "rate" here means take it per unit time.
LECTURE 9. φ4 THEORY & CROSS SECTIONS 77
Remark 9.2.1 . We call Γ the decay width because of the uncertainty principle: there is an
uncertainty in the energy of the particle, and so there is an uncertainty in the lifetime, and
so we do not get an exact value for the decay probability but a spread, or width.
Remark 9.2.2 . As we stated it, the sum in Equation (9.4) only accounts for final states that
are distinguishable (i.e. the number of particles and/or the type of particles change). However
we just saw in the previous section that certain processes come with symmetry factors, and
so we should slightly alter our expression to include these. That is we send
Γ → S · Γ,
Differential Probability 1 1
dσ := = |iMf i |2 dΠ, (9.6)
Unit Time × Unit Area 4E1 E2 V F
where E1 /E2 are, of course, the energies of our two initial state particles.
At first this looks a bit problematic because we still have a factor of V on the right-hand
side. However, we still need to insert the flux. So what is it? Well imagine that the beam is
moving with 3-velocity ~v perpendicular to some plane of area A. If the density of the particles
in the beam if ρ, then the total number of particles passing through the plane in time t is
simply
N
N = A|~v |ρt, =⇒ F = = |~v |ρ.
At
In the procedure above when we put the theory in a box, we did it in a normalisation where
there was exactly 1 particle in each volume V and so ρ = 1/V . So if we go to the centre of
mass frame of the collision, we have
|~v1 − ~v2 |
F = ,
V
where ~v1 /~v2 are the 3-velocities of our beams. Plugging this into Equation (9.6), we get
LECTURE 9. φ4 THEORY & CROSS SECTIONS 78
1 1
dσ = |iMf i |2 dΠ. (9.7)
4E1 E2 |~v1 − ~v2 |
Of course we get the full cross section by the same idea as for the decay rate: integrate of the
LIPS, sum over all the different final states and then multiply by the symmetry factor.
d3 p~1 d3 p~2
dΠ2 = (2π)4 δ (4) (p1 + p2 − pI ) .
(2π)3 2Ep~1 (2π)3 2Ep~2
The two d3 p~i s give us 6 degrees of freedom in the final state (i.e. the 3 components of each
final state particle), however the delta function kills 4 of these degrees of freedom leaving us
with just 2. These two degrees of freedom can be made very simple by going to the centre of
mass frame for the collision. Here we know that the final state particles must have equal and
opposite momentum, and so we can categorise the whole phase space by simply specifying
the modulus of the momentum and the angle at which these particles fly off relative to the
collision axis. In the diagram below, the red arrows are meant to be the final state particles.
|~
p|
∂(E1 + E2 ) −1
.
p1 |
∂|~
φ φ
p4
ψ ψ ψ ψ
y y
p2 p4 p2
7
I considered calling them p3 /p4 from the beginning but that might suggest we are considering the scattering
from the get go, but as the above remark says, this result also holds for the decay, so I decided against it.
LECTURE 9. φ4 THEORY & CROSS SECTIONS 80
p1 p4
p2 p3
which is known as an s diagram. Collectively these three types of diagrams describe all
possible (tree level, i.e. no loops) 2 → 2 scattering processes, and the variables s, t, u are
known as the Mandelstam variables. They are given by the momentum that flows through
the propagator,
s := (p1 + p2 )2 = (p3 + p4 )2
t := (p1 − p3 )2 = (p2 − p4 )2 (9.9)
u := (p1 − p4 )2 = (p2 − p3 )2
Exercise
Show that
4
X
s+u+t= m2i (9.10)
i=1
So why have we bothered doing this? Well firstly we can substitute these expressions
for t and u into our matrix elements and then take the complex norm squared to find our
differential cross section, Equation (9.8). However it allows us to see something else quite
interesting: because both t and u are negative we see the denominators of Equation (9.12)
never vanish. This is the statement that the propagator is never on shell (i.e. p2 = m2 never
occurs). We refer to such particles as virtual particles.
10 | Dirac Fields
So far all we have considered are Lorentz scalars, i.e. fields that transform as
under Lorentz transformations. We saw (it was set as an exercise in lecture 3, the one about
showing J i |~
p = 0i = 0) that quantised version of such fields have zero spin. Obviously we
know this is not the case in the real world, and we need to study other fields that transform
differently under Lorentz transformations. In particular we want to find something that once
we quantise it will give us spin-1/2 particles, as these are what electrons and the like are. It is
important to note that there is no a priori way to see what kind of transformation property we
would need to produce spin 1/2 particles, but we would have to undergo the full calculation
for a given transformation and then see what the resulting spin in. Luckily for us, however,
Dirac has basically done this for us and so we do know what transformation property we
require.
Remark 10.0.1 . This part of the course was forced to be a bit rushed (due to lack of time)
and so in order to try help expand on some of the points I am going to use Prof. Tong’s
notes as a heavy reference. I shall try outline most of the important points given in his notes,
however I don’t see the point in just copying his explanations out, so if there is anything I
say that sounds confusing, I strongly recommend having a look in his notes to see if there is
a more detailed explanation there.
As we have tried to make clear in these notes, Lorentz invariance is crucial to QFT and is a
vital part of the form we allow our Lagrangian to take. We have also just claimed that fields
that transform in different ways under the Lorentz transformation will give rise to different
types (namely different spins) of particle. In order to talk about how something transforms
under the action of a group we use a representation, i.e.1
00
φab... → D[Λ]a a0 D[Λ]b b0 ... φa b ... ,
where D[Λ] are matrices. The obvious question to ask is "what are the representations of the
Lorentz group?" This is the question we shall now try answer.
1
See my notes on Dr. Dorigoni’s Group Theory course for more details.
82
LECTURE 10. DIRAC FIELDS 83
where the ρ, σ ∈ {0, ..., 3} are antisymmetric, that is Ωρσ = −Ωσρ and the same for M. We
note that we still only have 6 basis elements, as ρ/σ can take 4 values, but the antisymmetry
gives us 4(4−1)/2 = 6 independent choices. This is of course a necessary condition (otherwise
we would be changing the dimension of our algebra!).
Remark 10.1.1 . It is worth clarifying the indices here. The ρσ values tell us which basis
element we are considering. Each Mρσ is a 4 × 4 matrix and the µν indices tell us which
element of this matrix we are considering. These two set of indices have nothing to do with
each other, despite them being of very similar form.
So how do we decide on what form the basis takes? The answer is we use the one that we
know gives us the answer we want and look super clever at the end for getting it correct. We
use
(Mρσ )µν = η ρµ η σν − η σµ η ρν ,
where on the right-hand side all the indices tell us the elements of the metric matrix, but
we should keep the above remark in mind about what the indices mean in terms of our Lie
algebra.
Now, looking at Equation (10.1), it is clear that we actually want to lower one of the µ/ν
indices so that we can see how it acts on our fields. Using η µν ηνσ = δ µ σ , this gives us
(Mρσ )µ ν = η ρµ δ σ ν − η σµ δ ρ ν , (10.3)
and so we have
1
ω µ ν = Ωρσ (Mρσ )µ ν , (10.4)
2
2
Or at least I think it’s mentioned in these notes, if not it’s mentioned now.
LECTURE 10. DIRAC FIELDS 84
Remark 10.1.2 . Note that by lowering one of the indices we have broken the antisymmetry
property of M. For examples see Prof. Tong’s notes, page 82.
Exercise
Show that Equations (10.3) and (10.4) imply that
ω µν = Ωµν . (10.5)
Exercise
Prove that the the above basis satisfy the Lie bracket (here just commutator) relation
and then use the answer to guide you. You want to end up with an expression that is
a matrix labelled by µν, i.e. something of the form (...)µ ν .
This construction has been done infinitesimally, we can recover some finite Lorentz trans-
formation by taking the exponential of the result, i.e.
1
Λ = exp Ωρσ M ρσ
. (10.7)
2
{γ µ , γ ν } = 2η µν 1n×n (10.8)
2 2
γ0 = 1n×n , and γi = −1n×n (10.9)
σi
12×2 0 0
0
γ = , and γ = i
, (10.10)
0 −12×2 −σ i 0
where {σ i }i=1,2,3 are the Pauli matrices. This is a representation of the Clifford algebra, and
we refer to this particular one as the Dirac basis.
Remark 10.1.3 . Note we can write Equation (10.10) in a more compact form using the tensor
product, namely
γ 0 = σ 3 ⊗ 12×2 , and γ j = iσ 2 ⊗ σ j .
This notation makes doing long manipulations with the gamma matrices easier. However in
this course we will not use this notation any further.
Remark 10.1.4 . The definition we have given here, Equation (10.10), differs from the one
given by Prof. Tong, who uses the so-called Weyl basis. The only difference between the two
is that in the Weyl basis we have
0 0 12×2
γ = .
12×2 0
The Dirac basis and Weyl basis are related by a simple equivalence transformation, and so
they are equivalent representations of our Clifford algebra. It turns out that the Dirac basis
is useful for studying massive particles, while the Weyl basis is useful for studying massless
particles. It is for this reason that we also refer to the Weyl basis as the chiral basis (its helps
to study the chirality of massless particles).
Now you would be very justified in asking what this has to do with the Lorentz group and
the content of the previous subsection. The answer is it turns out we can combine these gamma
matrices in such a way as to produce something which obeys Equation (10.6), and so form a
representation of the Lorentz group. Ok so what is it? Well the first thing we note is that it
has to contain products of 2 gamma matrices, as Mρσ has 2 indices. Perhaps the two most
natural things to try (given that we want to obtain a commutator relation, Equation (10.6)) is
the commutator and anticommutator. The anticommutator wont do because Equation (10.8)
shows us that the commutator of these will vanish, but what about the commutator? This
will turn out to work.
LECTURE 10. DIRAC FIELDS 86
Exercise
Show that
1 1 1
S ρσ := [γ ρ , γ σ ] = γ ρ γ σ − η ρσ . (10.11)
4 2 2
Hint: Consider the cases ρ = σ and ρ 6= σ separately and then use Equation (10.8) to
get the η ρσ term in the result.
Remark 10.1.5 . As with Mρσ above, for each ρσ value we have a 4 × 4 matrix (its a product
of gamma matrices), and so we need to introduce another two indices in order to label the
entries of these matrices. We shall use α, β = 1, 2, 3, 4 to label these.5
Claim 10.1.6 . The matrices S ρσ satisfy Equation (10.6) and so form a representation of the
Lorentz group.
Proof. See page 84 of Prof. Tong’s notes. (Or if you’re feeling up for it, give it a bash
yourself)
Remark 10.1.7 . Note that Equation (10.11) is satisfied at the level of the Clifford algebra.
That is, we do not need to give a specific representation to obtain it. This is obviously a
requirement if we want to use this result going forward.
10.1.3 Spinors
As we have said the S ρσ are 4 × 4 matrices that form a representation of the Lorentz trans-
formations. We have also said, being matrices they carry two indices, which we label by α
and β, and so we see that the object they act on carries one index. That is
where we include the [Λ] to help us distinguish this from the S ρσ s. Note that we have the
same numbers Ωρσ appearing in the exponential. This is just because we want the action of
this representation to apply the same transformation to the field ψ a as it does to x.
Remark 10.1.8 . Note that both S[Λ] and Λ are 4 × 4 matrices and it is entirely possible that
we haven’t constructed anything new. That is it could be true that S[Λ] is equivalent (in a
representation theory sense) to Λ. It turns out that we haven’t done this, and a proof of this
can be found in Prof. Tong’s notes, pages 85 and 86.
5
Note we do not start from 0 here. This is because we’re talking about the entries of a matrix, not a
spacetime index or a gamma index.
LECTURE 10. DIRAC FIELDS 87
Claim 10.1.9 . The representation S[Λ] is not completely unitary. In particular it it the boost
part of the Lorentz transformations that are not unitary.
Proof. From Equation (10.13), it follows that if S[Λ] was to be unitary then we would require
†
S ρσ = −S ρσ ,
i.e. it is antihermitian, as we take Ωρσ to be real numbers.6 However it follows from Equa-
tion (10.11) that this antihermitian condition would require
†
γργσ = γσ γρ,
for all ρ, σ = 0, 1, 2, 3. This is the statement that we require that all the gamma matrices
be hermitian or antihermitian. We cannot achieve this, though, as Equation (10.9) tells
us that γ 0 has real eigenvalues (i.e. hermitian) but the γ i have imaginary eigenvalues (i.e.
antihermitian).
Dirac wanted to try and obtain an equation of motion that was linear in derivatives. That is
he wanted something of the form
(Aµ ∂µ − m)ψ(x) = 0,
where the Aµ is something included in order to ensure a Lorentz invariant result. The question
is "what is Aµ ?" Well given that we’ve just spent a bunch of time talking about them, it might
seem sensible to suggest a gamma matrix, i.e. try something of the form
(γ µ ∂µ − m)ψ(x) = 0.
If we were to obtain this result, we would have to start from something which is bilinear in the
field ψ(x), as we differentiate one away in the Euler-Lagrange equations. We also note that
ψ(x) here is a column matrix (otherwise acting on it with the gamma matrix wouldn’t make
any sense),7 so if the action meant to just be a number (which it is) we also need a transpose
of ψ(x) appearing to the left of ψ in our action. Then we finally recall that it is QFT we want
to go on and study,8 so its best to be safe and turn this transpose into a hermitian conjugate.
Perhaps a better argument for this latter point is we want the result the action gives us to be
a real number. We therefore make the naïve guess
Lnaïve = ψ † γ µ ∂µ ψ − mψ † ψ.
6
They are the rotation angles/boost speeds.
7
Note for this reason we really should write m1 for the mass term. However it is standard to be a bit
sloppy with notation and just assume that people notice that the identity matrix is implicit.
8
For emphasis, what we’re doing right now is all classical. We will quantise this theory next lecture.
LECTURE 10. DIRAC FIELDS 88
We need this to be Lorentz invariant (as that was one of the rules for constructing La-
grangians), so we need to check this.
We have
†
ψ(x) → S[Λ]ψ Λ−1 x ψ † (x) → ψ † Λ−1 x S[Λ] ,
=⇒
and so the †
ψ † (x)ψ(x) → ψ † Λ−1 x S[Λ] S[Λ]ψ Λ−1 x ,
but we just saw that S[Λ] is not unitary, so this term is not invariant for all Lorentz trans-
formations. Hmm... so what do we do? Well let’s try and work out what went wrong and
use that to help us find the correct answer. As we said in the previous subsection, we can
pick a representation such that γ 0 is hermitian, (γ 0 )† = γ 0 and the γ i s are antihermitian,
(γ i )† = −γ i . From these, and (γ 0 )2 = 1 and the anticommutation relations, we conclude
†
γ0γµγ0 = γµ ,
Exercise
Use the above relation to show that
†
S[Λ] = γ 0 S[Λ]−1 γ 0 . (10.14)
Hint: Use Equation (10.13), expanding it out and using the fact that a matrix commuted
with itself.
With the result of the previous exercise in mind, we define the Dirac adjoint
and so we have
= ψ † Λ−1 x γ 0 ψ Λ−1
= ψ Λ−1 x ψ Λ−1 x ,
where on the second line we have inserted 1 = γ 0 γ 0 and then used Equation (10.14) on the
third line. So we see this bilinear is Lorentz invariant. This is exactly the kind of thing we
need for our Lagrangian, but we also need to check the term with a γ µ in it.
LECTURE 10. DIRAC FIELDS 89
Exercise
Show that γ µ satisfies −1
S[Λ] γ µ S[Λ] = Λµ ν γ µ . (10.16)
Remark: This was set as an exercise in the course notes, so I don’t want to type the
answer here. However if you get really stuck, Prof. Tong’s book might help...
Corollary 10.2.1. Equation (10.16) implies that ψ(x)γ µ ψ(x) transforms as a vector under
Lorentz transformations.
Proof. Just consider the transformation (suppressing x arguments for notational brevity):
−1
ψγ µ ψ → ψ S[Λ] γ µ S[Λ]ψ
= Λµ ν ψγ ν ψ,
where we have used the fact that Λµ ν for given values of µ, ν is just a number so can be
pushed through ψ. This is exactly the transformation property of a vector under Lorentz
transformations.
This Corollary tells us that we can treat the index on the gamma matrix in the bilinear
term above as if it was a Lorentz spacetime index. This therefore tells us that the term
ψγ µ ∂µ ψ in the Lagrangian will be Lorentz invariant. To summarise, the Dirac Lagrangian is
where the factor of i is included for a reason that will be explained shortly. Varying this
equation w.r.t. ψ(x) gives us the Dirac equation
iγ µ ∂µ − m ψ(x) = 0. (10.18)
But what about if we vary Equation (10.17) w.r.t. ψ(x)? Well this gives
Exercise
Take the hermitian conjugate of this expression and show it gives the Dirac equation
again.
Remark 10.2.2 . Ok that mysterious i, why is it there? Well the action for the Dirac field is
simply Z
S = d4 x ψ(x) iγ µ ∂µ − m ψ(x).
We require the action to be real, and so we need the integral to be the same under complex
conjugation. Well the result of the integral is just a number, so taking the transpose does
LECTURE 10. DIRAC FIELDS 90
nothing (it’s a 1 × 1 matrix), so we need the integrand to be hermitian. This is where the i
comes in. When we take the hermitian conjugate, we will get the derivative acting on ψ(x),
so we need to integrate by parts to get it to act on the ψ(x). This will come with a minus
sign, and so the complex conjugation of the i cancels this minus sign.
Notation. There is a common, and very handy, notation when considering the contraction of
gamma matrices, it is a slash. For example we write
∂/ := γ µ ∂µ .
We don’t just do this for derivatives, but for general gamma matrix contraction, for example
we use
µ
p
/ := γ pµ
for the contracted momentum. In this notation the Dirac equation reads
(i∂/ − m)ψ(x) = 0.
Now we can split the product of two gamma matrices into the symmetric and antisymmetric
parts,
1 1 1 1
γ µ γ ν = γ (µ γ ν) + γ [µ γ ν] = {γ µ , γ ν } + [γ µ , γ ν ],
2 2 2 2
where the second line follows simply from the definition of the commutator and anticommu-
tator. Now we know that ∂µ ∂ν is symmetric in its indices and its true that
A[µν] B(µν) = 0,
and so we are just left with the anticommutator term. Finally use Equation (10.8) to get
1 µν
2η ∂µ ∂ν + m ψ(x) = ∂ 2 + m2 ψ(x) = 0,
2
2
where the last line follows from the fact that the thing we started with being 0. This is exactly
the Klein-Gordan equation.
We have seen that the Dirac bilinear ψγ µ ψ is transforms as a scalar under Lorentz transfor-
mations. A question we could ask is "what if we put other matrices in between ψ and ψ?"
Well these are 4 × 4 matrices, so if we can find 16 linearly independent matrices are can study
LECTURE 10. DIRAC FIELDS 91
how a general matrix sandwiched between ψ/ψ transforms. We have 5 linearly independent
matrices given by {1, γ µ }, and the question is "can we get the other 11 using only the gam-
mas?" The answer is yes. We have already found 6 others in S ρσ , so so far we have 11 out of
16. What else could we do? Well we could multiply all 4 of the gamma matrices together:9
γ 5 := iγ 0 γ 1 γ 2 γ 3 . (10.19)
This is actually a very important combination, and we shall discuss it further in a moment.
In the Dirac basis this takes the form
0 12×2
5
γ = . (10.21)
12×2 0
Remark 10.2.3 . In the Weyl/chiral basis we get
5 12×2 0
γ = .
0 −12×2
Comparing these two expressions with our definition of γ 0 (and recalling Remark 10.1.4) we
see that the essential different between the Dirac and Weyl basis is γ 0 ←→ γ 5 .
Equation (10.21) shows us that γ 5 is linearly independent from the other matrices in the
Dirac basis (and indeed this is true in an arbitrary representation), so we have 5 + 6 + 1 = 12
of the 16. What about the remaining 4? Well so far we have considered products of 2 gammas
(S ρσ ) and 2 gammas (γ 5 ), what about a product of three?
Exercise
Show that
γµγν γρ ∼ γσ γ5
for µ 6= ν 6= ρ 6= σ. Hint: Multiply the left-hand side by (γ ρ )2 ∼ 1.
The result on the right-hand side of the above exercise has a single spacetime index and so
corresponds to the final 4 linearly independent matrices.10 So we can write any 4 × 4 matrix
as a linear combination of the set
Γ = {1, γ µ , γ 5 , γ 5 γ µ , S ρσ }.
So we can write a general Dirac bilinear in the form ψΓψ. Extending the argument made
after Corollary 10.2.1, about treating the gamma index as a spacetime index, we can conclude
that these things transform as follows
9
Note we don’t call it γ 4 , despite it being the fourth gamma. This is just because in Euclidean space we
would have µ = 1, 2, 3, 4 and so 5 would make sense there.
10
We haven’t actually showed this is linearly independent, but you can check it in the Dirac and Weyl bases
at least.
LECTURE 10. DIRAC FIELDS 92
ψψ Scalar
ψγ µ ψ Vector
ψS µν ψ Tensor
ψγ 5 ψ ...Scalar
ψγ 5 γ µ ψ ... Vector
As you probably noticed, the last two terms have ellipses before them, the reason why is set
as an exercise.11
Exercise
Parity.
Parity is the transformation xµ → x0µ = (x0 , −~x), which can be thought of as mapping
the world onto its mirror image.
(a) Show that is ψ(x) is a solution to the Dirac equation, then so is ψ 0 (x0 ) := γ 0 ψ(x)
in the parity transformed world. In other words, start with (iγ µ ∂µ − m)ψ(x) = 0,
and manipulate it into the form
(b) Compute the transformation laws for the following bilinears under parity, i.e.
calculate
0
ψ(x)ψ(x) → ψ (x0 )ψ 0 (x0 ) =?
0
ψ(x)γ 5 ψ(x) → ψ (x0 )γ 5 ψ 0 (x0 ) =?
0
ψ(x)γ µ ψ(x) → ψ (x0 )γ µ ψ 0 (x0 ) =?
0
ψ(x)γ µ γ 5 ψ(x) → ψ (x0 )γ µ γ 5 ψ 0 (x0 ) =?
You should find that the bilinears transform as a scalar, pseudoscalar, vector and
axial vector, respectively.
Remark 10.2.4 . The last sentence in the exercise above is exactly the reason we put the ellipses
in the table above for the Lorentz transformation of Dirac bilinears.
σi
0 1 0
0
γ = and γ = i
.
1 0 −σ i 0
11
This was an exercise from the course, so I don’t want to type the results here. As always feel free to ask
me any questions for further clarity.
LECTURE 10. DIRAC FIELDS 93
where the R/B stand for rotations and boosts respctively, and where the notation ϕ ~ · ~σ =
ϕ σ + ϕ σ + ϕ σ and similarly for χ
1 1 2 2 3 3 ~ · ~σ . The numbers ϕ /χ are the parameters that tell
i i
where each ψR/L is a 2-component column matrix, we call these Weyl/Chiral spinors. Note,
from Equation (10.22) we see that ψR and ψL transform the same under rotations but oppo-
sitely under boosts!
We can actually use γ 5 to project our the ψR /ψL part of the field. We define
1 1
1 − γ5 , and PR := 1 + γ5 , (10.24)
PL :=
2 2
so that
ψL = PL ψ, and ψR = PR ψ.
We can see this is true for the Weyl basis explicitly, but we can actually define ψL/R in a
general representation by the above formulas. We call these the left-handed and right-handed
spinors (hence the subscripts). We will see why this is the case in the following exercise.
Remark 10.2.5 . Again the understanding of the remaining part of this material was set as an
exercise on the course, so I have just inserted the question here.
12
For more info see, for example, my notes on Dr. Dorigoni’s Group theory course.
LECTURE 10. DIRAC FIELDS 94
Exercise
(a) Use Equation (10.24) to show that PL and PR are projection operators, e.g.
PL2 = PL , PR2 = PR and PL PR = 0.
(c) Show how ψL and ψR look in the parity transformed world (see previous exercise).
That is perform the transformation and then express everything in terms of ψL0 (x0 )
and ψR0 (x). Show that the Lagrangian is invariant under parity transformation.
(d) Consider the Weyl spinor equation, Equation (10.23). Set m = 0 and write out
the explicit form of the Dirac equation in this basis.
(e) Plug in plane wave solutions ψ(x) = u(p)e−ip·x into the massless Dirac equation
(note p0 = |~p| when m = 0) and show that uL and uR are eigenvectors of the
helicity operator
1 p̂ · ~σ 0
h := .
2 0 p̂ · ~σ
Here p̂ is a unit vector in the direction of p, which you can always choose to be,
say, in the z-direction. List the eigenvalues of the left and right-handed spinors.
(After quantisation these correspond to the situation where the particle spin is
either aligned or anti-aligned with the direction of motion.)
Remark 10.2.6 . Part (e) of the exercise above is the motivation for the names left/right-
handed. The result should say that the helicity of ψL and ψR are opposite. Helicity basically
tells you the projection of the spin of a massless particle onto its momentum, we call the
two options left- and right-handed. These names come from our hands: make a thumbs up
but don’t curl your fingers all the way in, now imagine your thumb points in the direction of
momentum, then your fingers tell you about the spin direction. A right-handed spinor has
spin-momentum projection like your right hand looks, and similarly for a left-handed spinor.
Left-Handed Right-Handed
Remark 10.2.7 . As a final remark to this lecture, we started it by saying that we were going to
find quantum fields corresponding to spin-1/2 particles. To emphasise again, it is not trivial
to see at this stage that this is what we have done. We will return to this next lecture and
compute the angular momentum and show we do indeed have spin-1/2 particles.
11 | Quantising The Dirac Field
We now want to proceed and try and quantise our Dirac Lagrangian and arrive at some QFT.
In order to do this, we are going to draw comparisons to the complex scalar field and use the
results from that to guide us.
For the complex scalar field we could expand the classical theory as an integral over Fourier
modes:
d3 p~ 1 −ip·x
Z
† ip·x
φ(x) = bp~ e + cp~ e .
(2π)3 Ep~
p 0
p =Ep~
When we quantised this, promoting φ(x) to an operator, we saw it was built out of the classical
plane wave solutions to the Klein-Gordan equation with the creation/annihilation operators
being the Fourier coefficients. We now want to start from a similar anstaz for the Dirac field,
but now we need to take into account the fact that the Dirac equation is a matrix equation
(it has gamma matrices in it) and so we need to include some 4-component column matrices
in there too. So we propose the anstaz
d3 p~ 1
Z
−ip·x † ip·x
ψ(x) u(~
p , s)bp~ e + v(~
p , s)cp~ e . (11.1)
(2π)3 Ep~
p 0
p =Ep~
where, as we will see, the s will label the different spin states, i.e. s = 1, 2. We then define
ψ(x) from this, giving us
d3 p~ 1
Z
−ip·x † ip·x
ψ(x) u(~
p , s)cp~ e + v(~
p , s)bp~ e . (11.2)
(2π)3 Ep~
p 0
p =Ep~
Exercise
Use the Dirac equation, Equation (10.18), to show that
/ − m)u(~
(p p, s) = 0, and (p
/ + m)v(~
p, s) = 0 (11.3)
95
LECTURE 11. QUANTISING THE DIRAC FIELD 96
We can use the result of this exercise to obtain the explicit form of u(~
p, s) and v(~
p, s) in
a given representation. We will choose to use the Dirac basis. We also note that the Dirac
equation is Lorentz invariant and so we can pick whichever frame we like, and so we choose
the rest frame. We therefore get
0
p
/ = mγ , 0, 0, 0),
and so
0
0 0
0 = m γ − 14×4 u(~
p, s) =
0 12×2
so we get two independent solutions, i.e. the two s values:
1 0
√ 0 √ 1
u(~
p, 1) = 2m 0 , and u(~ p, 2) = 2m
0 .
0 0
√
where we have normalised the states with 2m. Similarly we get
0 0
√ 0 √ 0
v(~p, 1) = 2m 1 ,
and u(~
p, 2) = 2m
0 .
0 1
√
Remark 11.1.1 . A nice justification for the fact that we normalised the states with a 2m
above can be found in Prof. Tong’s notes on page 100-102. However, it is important to note
that he is using (and so does Peskin and Schroeder) the Weyl basis. This is related to the
Dirac basis via the similarity transformation
1 12×2 −12×2
S=√ ,
2 12×2 12×2
0 1
Claim 11.1.2 . The normalisation given for our u/v results in the normalisation
u(~
p, s)u(~
p, s) = 2m, and v(~ p, s) = −2m
p, s)v(~
Proof. We do this by noting that u(~ p, s) transforms the same as ψ(x) and so we know that
the products u(~ p, s) and u(~
p, s)u(~ p, s)γ µ pµ u(~p, s) are Lorentz invariant so for the following
expression
u(~
p, s)p
/u(~ p, s) − mu(~ p, s) = 0
LECTURE 11. QUANTISING THE DIRAC FIELD 97
u(~ p, r) = 2mδ rs ,
p, s)u(~ and v(~ p, r) = −2mδ rs .
p, s)v(~ (11.4)
It is important to note that this result is Lorentz invariant and basis independent and so will
hold in general.
2
X 2
X
u(~
p, s)u(~
p, s) = p
/ + m, and v(~
p, s)v(~ / − m,
p, s) = p (11.5)
s=1 s=1
where the two spinors are not contracted but placed back to back to give a 4 × 4 matrix.
Proof. The easiest way to prove this, we would be to go back and look at the solution of
Equation (11.3) in the Weyl basis and obtain an expression for u(~ p, s) and v(~
p, s) and then
use these results to show Equation (11.5). These are done in Prof. Tong’s notes on pages
100-101 ad 104-105, respectively, and I don’t see the point in just copying this out when I can
just point the reader there.
11.2.1 Fermions
The study of the atomic spectra of fermions showed that fermionic wavefunctions are antisym-
metric under exchange of quantum numbers. It was shown by Jordan and Wigner that this is
equivalent to creation/annihilation operators obeying anticommutator relations, in contrast
to the commutator relations in the case of bosons. Let’s say this matematically.
LECTURE 11. QUANTISING THE DIRAC FIELD 98
Our states are defined in the same way to the bosonic case, i.e.
where α is some general quantum number, with multiparticle states then given by
In the bosonic case it didn’t matter what way we ordered the creation operators as they
commuted, however, as we have just said, Fermi statistics tell us that the wavefunction changes
sign if we switch the quantum numbers α ↔ β, i.e. we require
|α, βi = − |β, αi .
{bα , bβ } = 0.
Now note that the antisymmetric nature tells us that we cannot create two particles with the
same quantum numbers, and the same for annihiliations i.e.
This essentially tells us we can label the elements of our Hilbert space by stating whether
the slot is filled or not, e.g. |0, 0i and |1, 0i etc. From this we can write the action of the
creation/annihilation operators as
The fact that this is true is meant to be motivated by the next exercise.
Exercise
Show that for a system that only contains two particles, i.e. the only states available are
|0, 0i , |0, 1i , |1, 0i and |1, 1i, that {bα , b†β } = δαβ for α, β = 1, 2. Hint: You will want to
use the antisymmetric property of the states to make statements like |0, 1i + |1, 0i = 0.
LECTURE 11. QUANTISING THE DIRAC FIELD 99
Exercise
Show that Equation (11.6) is equivalent to
and all other anticommutators vanishing. Hint: Use Equations (11.1) and (11.2) and
then use (γ 0 )2 = 1 so that you can use the spin sums Equation (11.5). If you cant work
it out, Prof. Tong does it for commutation relations in his notes on page 107 which
should give you the outline of the proof.
Remark 11.2.1 . Note it is easy to see from the above exercise that
ψ α (~x), ψ β (~y ) = 0
as we see in Equation (11.2) that we will only get anticommutators between c and b† .
Remark 11.2.2 . Note it is not a fully convincing argument to say "the quantisation of Dirac
fields are spin-1/2 particles and we know spin-1/2 particles are Fermions so we should use
Fermi statistics." Firstly, we haven’t yet shown that Dirac fields correspond to spin-1/2 par-
ticles. We actually could have done this already, but even still I don’t believe its a convincing
argument to say we therefore use Fermi statistics for the fields. As we will see there is a much
more physically crucial1 reason why we have to use anticommutation relations.
π = iψγ 0 = iψ † γ 0 γ 0 = iψ † , (11.8)
H = ψ − iγ i ∂i + m ψ.
This is the classical result, we get the quantum version by promoting everything to operators
and inserting our operator expressions Equations (11.1) and (11.2). Doing this we obtain2
2
d3 p~ X †
Z
†
H= bp~,s bp~,s − cp~,s cp~,s Ep
~
(2π)3
s=1
2
d3 p~
Z
bp†~,s bp~,s + cp†~,s cp~,s − (2π)3 δ (3) (0) Ep~
X
= (11.9)
(2π)3
s=1
2
d3 p~
Z
bp†~,s bp~,s + cp†~,s cp~,s Ep~
X
∴ H=
(2π)3
s=1
where we have used the anticommutation relation Equation (11.7), and where the therefore
symbol contains the information about us dropping the infinity term by a redefinition of the
ground state energy, as we did for the scalar case.
So we see that the theory contains two types of particles, one created by b† and the other
by c† . These particles have the same mass (as they give the same energy contribution, Ep~ ,
and each carry one of two spins. This is exactly what we expect of spin-1/2 Fermions, and
we shall indeed provide further support of this claim. We use this interpration to write the
Hamiltonian in terms of the number operators
2
d3 p~ X
Z
H= Nb,s + Nc,s Ep~ .
(2π)3
s=1
Remark 11.2.3 . Notice that in Equation (11.9) the ground state energy has the opposite sign
to the scalar case. That is for the scalar field we have the ground state energy being +∞
while here we have it being −∞.
2
To show this we need the same relations we need for the proof of the spin sums, Equation (11.5). Seeing
as these aren’t presented here, we do not prove this formula. Again it is done in Prof. Tong’s notes, page 108.
Note to self: Maybe after the courses are finished and you have more time come back and include all this stuff.
LECTURE 11. QUANTISING THE DIRAC FIELD 101
Remark 11.2.4 . We can now wee why we needed to consider anticommutation relations and
not commutation relations. If we had used the commutation relation
then when we switched the cp~,s cp†~,s term in the first line of Equation (11.9) we would obtain
2
d3 p~ X †
Z
† 3 (3)
H= bp~,s bp~,s − cp~,s cp~,s + (2π) δ (0) Ep~ .
(2π)3
s=1
Again we can drop the delta function however the negative sign infront of the cc† term is
a dagger3 to the heart of our theory; it tells us that we our spectrum is unbounded from
below! That is we could continually lower the energy of our system by repeated creation
of c† particles. This is obviously not good, and it is for this reason that we had to use
anticommutation relations for our fields.
T µν = iψγ µ ∂ ν ψ − η µν L
= iψγ µ ∂ ν ψ − η µν ψ(i∂/ − m)ψ.
This doesn’t look particularly nice because of the second term. However we can make this
much nicer by noting that the a current is conserved only when the equations of motion
are satisfied, so we can freely impose Equation (10.18) to set the second term to zero. We
therefore have
T µν = iψγ µ ∂ ν ψ. (11.10)
From this we can find the physical momentum via
Z
P := d3 ~x T 0i
i
2
d3 p~ X †
Z
†
= p
~ bp~,s bp~,s + cp~,s cp~,s
(2π)3
s=1
2
d3 p~
Z X
= p~ Nb,s + Nc,s ,
(2π)3
s=1
which just tells us that the total momentum (not including spin!) in the system is given by
the sum of momentum from the b† particles and the c† particles.
3
Awful pun, I know...
LECTURE 11. QUANTISING THE DIRAC FIELD 102
Remark 11.2.5 . Note that we could have also used the equations of motion when considering
the scalar fields earlier in the course, however they weren’t really any help because the stress-
tensor was linear in derivatives whereas the equations of motion were quadratic. This is one
of the beauties of the Dirac equation being linear in derivatives.
Now let’s look for the conserved charge corresponding to the internal U (1) symmetry:
ψ → e−iα ψ
for some fixed α ∈ R. For the scalar field case this gave rise to conservation of particle number
for the real scalar field and conservation of the difference of particle numbers for the complex
scalar field. We can show this gives rise to the current
j µ = ψγ µ ψ,
It is at this point that we get our particle antiparticle interpretation. This is because of the
relative minus sign above which tells us that the c† particle carries the opposite charge to the
b† particle.
Exercise
∇µ j µ = 0.
11.2.5 Spectrum
We can now look at the spectrum of the system. As always we define the vacuum to be the
state that is annihilated by all annihilation operators,
for all p~ and s. It follows immediately from the previous expressions that the vacuum obeys
which tells us the vacuum has zero energy, zero momentum and no particles.4
4
Note the condition Q |0i = 0 only tells us that there is an equal number of particles and antiparticles, but
because the momentum and energy are zero, it follows that we cannot have any particles.
LECTURE 11. QUANTISING THE DIRAC FIELD 103
which again supports our statement that b† makes particles while c† makes antiparticles.
Exercise
Find the energy, momentum and charge of multiparticle states. Hint: Be careful about
the states you allow, as we have Fermions so |~
p, s; p~, siψ = 0 by antisymmetry, and
similarly for the antiparticle states.
11.2.6 Spin
So what about the spin? Well, recalling what we did in lectures 2 and 3, we get this by
considering an infinitesimal Lorentz transformation on our system. We have already seen,
Equation (10.12), that the Dirac spinor transforms under the Lorentz group as
ψ α (x) → S[Λ]α β ψ β Λ−1 x ,
so we just need to consider the infinitesimal version of this. We have already seen
Λµ ν ≈ δ µ ν + ω µ ν
and from Equation (10.13) we have
1 α
S[Λ]α β ≈ δ α β + Ωρσ S ρσ β ,
2
which gives us
1 α
δψ α (x) = −ω µ ν xν ∂µ ψ α (x) + Ωρσ S ρσ β ψ β (x),
2
where the minus sign on the first term comes from the fact that we have Λ−1 .5 Now we use
Equation (10.5), which says
ω µν = Ωµν ,
to obtain h 1 α i
δψ α (x) = −ω µν xν ∂µ ψ α (x) − S µν β ψ β (x) .
2
If we compare this to the derivation of Equation (2.12) we see that the first term above is
exactly the same, while the second term above is new so will give an additional contribution
to the Lorentz current. In total we have
(J µ )ρσ = xρ T µσ − xσ T µρ − iψγ µ S ρσ ψ.
5
See Section 2.1.1 if this doesn’t make sense.
LECTURE 11. QUANTISING THE DIRAC FIELD 104
where6 k
~ = S 23 , S 31 , S 12 , i σ 0
S ij = − ijk
S ,
2 0 σk
where σ k are the Pauli matrices. This expression can be used to show that our particles do
indeed have spin-1/2. We do not do the calculation here but instead point the interested
reader to pages 60-61 of Peskin and Schroeder.
Exercise
Show the Equation (11.13) holds. Hint: Do it for one component J i and notice how it
extends to this formula.
12.1 Propagators
12.1.1 Causality
So far we have quantised the Dirac equation and showed (or at least claimed it can be shown)
that they correspond to spin-1/2 Fermions. The next thing we need to check is causality, that
is we need the field operators to not communicate outside the lightcone. For the complex
scalar field we saw that this boiled down to the condition
However there is a problem here: we can expand ψ/ψ in terms of the creation and annihilation
operators b, b† , c and c† , however we do not know what they commutation relationships are.
That is
[bp~,s , b†q~,r ] =?
Now we can relate these to the anticommutation relations using the identity
which is easily verified. However this raises a problem as, using Equations (11.1) and (11.2),
we see that the commutator will contain terms that have creation operators to the right. For
example
[cp†~,s cq~,r ] = {cp†~,s cq~,r } − 2cq~,r c†q~,s ,
and while the first term will just give us a delta function, as per Equation (11.7), the second
term will give a non-zero expectation value. Equally we will get terms arising containing both
b and c operators, e.g.
[cp†~,s , b†q~,r ] = −2b†q~,r cp†~,s ,
where we have used the fact that the anticommutator vanishes. It is not trivial to see that
all these contirbutions will somehow cancel and give us a vanishing commutator for spacelike
separations.
105
LECTURE 12. DIRAC PROPAGATORS & PERTURBATION THEORY 106
So what do we do? Well we note that every single observable quantity we have written
for the Dirac field has contained bilinears, ψΓψ. Note this is obviously true because what we
measure are numbers but ψ/ψ are matrices, so we need to contract them to get a number.
We therefore propose that what we need for our causality condition is
h i
ψ(x)ψ(x), ψ(y)ψ(y) = 0 ∀(x − y)2 < 0. (12.2)
Proof. We just need to show the two relations are related, and then of course they agree under
the given conditions. That is we need to show
h i
ψ(x)ψ(x), ψ(y)ψ(y) ∼ ψα (x), ψ β (y) .
[AB, C] = A[B, C] + [A, C]B, and [A, CD] = C[A, D] + [A, C]D.
We will also need to use the anticommutation relations Equation (11.6) along with
ψψ = ψ α ψα ,
where there is an implicit sum. To save notation we shall use α for the fields as a function of
x and β for the y fields, we therefore have
h i h i h i
ψ α ψα , ψ β ψβ = ψ α ψα , ψ β ψβ + ψ α , ψ β ψβ ψα
h i h i h i h i
= ψ α ψα , ψ β ψβ + ψ α ψ β ψα , ψβ + ψ α , ψ β ψβ ψα + ψ β ψ α , ψβ ψα
= ψ α ψα , ψ β ψβ −2ψ α ψ β ψα ψβ + ψ α ψ β ψα , ψβ −2ψ α ψ β ψβ ψα
+ ψ α , ψ β ψβ ψα − 2ψ β ψ α ψβ ψα + ψ β ψ α , ψβ ψα − 2ψ β ψβ ψ α ψα
= ψ α ψα , ψ β ψβ + ψ β ψ α , ψβ ψα −2ψ α ψ β ψα , ψβ } − 2ψ β ψ α ψβ ψα
−2ψ β ψβ ψ α ψα
= ψ α ψα , ψ β ψβ + ψ β ψ α , ψβ ψα − 2ψ β ψ α ψβ ψα −2ψ β ψβ , ψ α ψα
+2ψ β ψ α ψβ ψα
= ψ α ψα , ψ β ψβ + ψ β ψβ , ψ α ψα − 2ψ β ψβ , ψ α ψα
= ψ α ψα , ψ β ψβ − ψ β ψβ , ψ α ψα ,
where I have tried to colour/underline the terms that carry over from line to line. For clarity,
we have used
ψα , ψβ } = 0 = ψ α , ψ β }
to drop all terms of that form.
LECTURE 12. DIRAC PROPAGATORS & PERTURBATION THEORY 107
Ok so we can impose our causality condition by requiring Equation (12.3) holds. This is
left as an exercise.
Exercise
Show that
(12.4)
iSαβ (x − y) = i∂/ x + m αβ
D(x − y) − D(y − x) ,
where D(x, y) is the propagator from the scalar theory, i.e.
d3 p~ 1 −ip·(x−y)
Z
D(x − y) = e .
(2π)3 2Ep~
Hint: You will want to use the spin sums, Equation (11.5). If you get stuck, Prof.
Tong does this calculation on page 113.
Remark 12.1.2 . Recall that the propagator D(x − y) is in the kernel of the Klein-Gordan
equation if we are on shell, p2 = m2 , that is
Remark 12.1.3 . For some of the expressions that follow, we shall drop the αβ labels for the
matrices, however it is important to remember that our results are matrices. This should
generally be clear given that gamma matrices will appear.
: bp~,s bq~,r b~† : = (−1)2 : b~† bp~,s bq~,r : = b~† bp~,s bq~,r .
`,t `,t `,t
where we note the minus sign in the cases. This minus sign is necessary to preserve Lorentz
invariance: recall that when (x − y)2 < 0 there is no Lorentz invariant way to decided if
x0 > y 0 or y 0 > x0 , putting this together with Equation (12.3) we see the minus sign is
needed. That is, Equation (12.3) tells us
ψ(x)ψ(y) = −ψ(y)ψ(x),
for spacelike separation, but in here we can transition from x0 > y 0 to y 0 > x0 using a Lorentz
transformation, so we need a minus sign in the time ordering expression.
Exercise
Use the Fourier expansions Equations (11.1) and (11.2) to show that
d3 p~ 1 −ip·(x−y)
Z
h0| ψα (x)ψ β (y) |0i = 3
e p
/ + m αβ
(2π) 2Ep~
d3 p~ 1 ip·(x−y)
Z
− h0| ψ β (y)ψα (x) |0i = − 3
e / − m αβ .
p
(2π) 2Ep~
Hint: If you have done the previous exercise this one is basically done.
Claim 12.1.6 . We can combine the above two relations to write the Feynman propagator as
a 4-momentum integral
d4 p i(p
/ + m) −ip·(x−y)
Z
SF (x − y) = e . (12.6)
(2π) p − m2 + i
4 2
Proof. You do a contour integral. The calculation is similar to the one we did for the Feynman
propagator for the scalar field, Equation (5.10). We ‘prove’ it here by noting the similarity to
that expression.
Remark 12.1.7 . Note in extension of Remark 12.1.2, we see that the Feynman propagator for
Ferimons is a Green’s function of the Dirac equation, that is
i∂/ x − m SF (x − y) = iδ (4) (x − y).
LECTURE 12. DIRAC PROPAGATORS & PERTURBATION THEORY 109
We now want to look at the interacting theory and derive the Feynman rules. As before we do
this by using Wick’s theorem and considering the contractions. As we mentioned above, we
need to make a small tweak to Wick contractions in order to account for the anticommutation
relations.
Then if we are taking a contraction within a string of normal ordered objects we obtain the
result by using the anticommutation relations to put the contracted fields next to each other
and then normal ordering the rest.
Example 12.2.1 . Let’s say we want to contract ψ1 with ψ5 and ψ2 with ψ3 in the following
normal ordering. Then we move the ψ5 so that it’s next to the ψ1 and remove the two
contractions. That is
: ψ1 ψ2 ψ 3 ψ4 ψ 5 ψ6 : = (−1)3 : ψ1 ψ 5 ψ2 ψ 3 ψ4 ψ6 :
= −SF (x1 − x5 )SF (x2 − x3 ) : ψ4 ψ6 :
We then define the contraction of the operators on states. However we need to account
for the fact that we have two different kinds of states, and so in total we have 4 kinds of
contractions. These correspond to the two types of particle (i.e. particle and antiparticle)
and whether they are ingoing or outgoing. We list them in the table below.
There’s four things to notice here, which we list in bullet points below.
1
See Section 7.2 if you need a reminder on how we obtain this.
LECTURE 12. DIRAC PROPAGATORS & PERTURBATION THEORY 110
• We only contract ψ with the ψ states (i.e. have a subscript ψ). Similarly for contracting
ψ.
• Incoming particles come with a factor of e−ip·x while outgoing particles come with e+ip·x .
• The particles are given by the letter u while the antiparticles are given by v.
• If the contraction is obtained using a ψ then the resulting symbol has a bar over it, e.g.
an outgoing particle has a u.
Remark 12.2.2 . Note that if we had a multiparticle state, as per the method of putting the
contracted objects next to each other, we need to switch the ordering of the ket so that we
contract the field with the first entry. With the Fermion states discussion above, we see this
gives us potential minus signs. For clarity, if we want to contract ψ(x) with the second entry
in a two particle state, we have
ψ(y)ψ(x) |~
p, s; ~q, ri = −ψ(x)ψ(y) |~
p, s; ~q, ri
Yakawa theory carries over2 to the Dirac fields, and the Lagrangian is given by
1 1
L = ψ i∂/ − m ψ + (∂φ)2 − µ2 φ2 − gψψφ,
2 2
where µ2 is the mass term for the real scalar field φ.
Before we even derive the Feynman rules we can sort of guess what the diagrams will look
like. We see that we have three fields ψ, ψ and φ and the interaction term couples one of each
together with coupling strength g. So we expect the diagram to have a vertex something like
ψ
p~1
~k
x
φ
p~2
ψ
2
In fact this is the actual Yakawa theory. The scalar Yakawa theory was just studied because its useful for
the study now.
LECTURE 12. DIRAC PROPAGATORS & PERTURBATION THEORY 111
This represents the process ψ(p1 )ψ(p2 ) → φ(k), which is particle-antiparticle annihilation.
Now it follows from the fact that we can contract the ψ with an outgoing state and produce
a particle, that we can rotate this Feynman diagram and obtain the following one:
φ
p~1
~k
x
ψ
p~2
ψ
which represents the scattering process φ(p1 )ψ(p2 ) → ψ(k). Note there are no antiparticles
present here. Indeed it we can always rotate a Feynman diagram and obtain new ones,
provided the particles in the new one are valid. That is, both of the above diagrams are fine
because we have particles and antiparticles in this system.3 Note that we always have one
Fermion arrow (i.e. the arrows on the lines) flowing into a vertex and one flowing out. This
corresponds to our charge Equation (11.11) being conserved locally at the vertex.
Ok let’s find some Feynman rules. We’ll start by looking at the scattering process
There are two distinct types of contractions we can do here. We shall use our guess work
above to draw these diagrams now for pictorial clarity, however it is important to note that
it could be entirely possible that what we draw doesn’t correspond to the Wick contraction
result at all.4 The two diagrams are:
p~1
ψ ψ x p~3
p~1 p~3 ψ ψ
y x ψ
ψ
p~2 p~4 φ p~4 φ
φ φ y
p~2
Let’s just focus on the left-hand one. This corresponds to a second order term in Dyson’s
formula and is given by the following contraction (note we have included a α/β index on the
ψ/ψs to show how the matrices are contracted. Also note that we have used |ψ(p, s)i instead
of |p, siψ to help keep track of what comes from where.)
(−ig)2
Z
α β
d4 xd4 y hφ(p4 )ψ(p3 , s)| ψ x ψxα φx ψ y ψyβ φy |ψ(p1 , r)φ(p2 )i ,
2!
3
I have never actually met a case where this doesn’t work out, but I thought it was worth clarifying in case
such cases do exist.
4
Of course it will, but I just mean to highlight that just because you think it should look like this, doesn’t
mean it has to.
LECTURE 12. DIRAC PROPAGATORS & PERTURBATION THEORY 112
(−ig)2
Z
d4 x d4 y ei(p3 +p4 )·x u(~ p1 , r)β e−i(p1 +p2 )·y .
p3 , s)α SF (x − y)αβ u(~
2!
We can then insert the integral expression for SF (x − y), Equation (12.6), as
d4 q i(/q + m) −iq·(x−y)
Z
SF (x − y) = e ,
(2π) q − m2 + i
4 2
and then do the integral over y and q to set q = p1 + p2 , and then the integral over x gives a
delta function δ (4) (p1 + p2 − p3 − p4 ), which gives us the expression
(−ig)2 (4) i p/1 + p
/2 + m αβ
δ (p1 + p2 − p3 − p4 )u(~ p3 , s)α u(~
p1 , r)β .
2! (p1 + p2 )2 − m2 + i
Hint: Note the minus sign in the numerator, this should come from a delta function.
The results of the above exercises suggest the following momentum space Feynman rules.
All the usual stuff for the scalar field φ and then:
LECTURE 12. DIRAC PROPAGATORS & PERTURBATION THEORY 113
q Z
d4 q i(/q + m)
Fermion Propagator
(2π)4 q 2 − m2 + i
q Z
d4 q i(−/q + m)
Antifermion Propagator
(2π)4 q 2 − m2 + i
−ig
Vertex −(ig)
Note that we also label the spin s on the momentum arrow, and note again that the antifermion
propagator comes with a −/q in the fraction.
Exercise
Draw the corresponding Feynman diagram to the previous exercise.
k and k
ψ ψ ψ ψ
y y p~4 , s4
p~2 , s2 p~4 , s4 p~2 , s2
The relevant term from Dyson’s formula is (forgetting about the integrals etc, its just the
contractions we care about here)
hψ(~ p3 , s3 )| ψ x ψx φx ψ y ψy φy |ψ(~
p4 , s4 )ψ(~ p1 , s1 )ψ(~
p2 , s2 )i
5
Note I have tried to colour coordinate the momentum/spin to make it more clear.
6
See QED lectures for more information.
LECTURE 12. DIRAC PROPAGATORS & PERTURBATION THEORY 115
Now if we want to only have Fermions and no antifermions, we will need to contract the ψs
with the bra vectors and the ψs with the ket vector. So firstly we move them, picking up the
minus signs. Step by step, this gives us
− hψ(~ p3 , s3 )| ψ x ψ y ψx φx ψy φy |ψ(~
p4 , s4 )ψ(~ p1 , s1 )ψ(~
p2 , s2 )i
= − hψ(~ p3 , s3 )| ψ x ψ y φx φy ψx ψy |ψ(~
p4 , s4 )ψ(~ p1 , s1 )ψ(~
p2 , s2 )i
were we have use the fact that we can move the ψ through the φs with no problems. We now
use Remark 12.2.2, which basically tells us that we have to order our ψx ψy /ψ x ψ y so that we
have the correct momenta/spin going to the correct vertex, i.e. x or y. The convention is we
fix what we call x and what we call y by where the initial momentum flows.7 This means
we essentially don’t have a choice for the ψx ψy contractions. As per the diagrams above, we
choose to have ψ(~p1 , s1 ) flowing into vertex x,8 so we need to move the ψx closest to the ket
vector. Clearly we also have to contract the φx and the φy to get the propagator between the
vertices. So we get
hψ(~ p3 , s3 )| ψ x ψ y φx φy ψy ψx |ψ(~
p4 , s4 )ψ(~ p1 , s1 )ψ(~
p2 , s2 )i
for both the t and u channel. Seeing as this is the same for both we can drop it here and
just look at the contractions between the ψs and the bra-vector. For the t channel we need
to contract the ψ x with the ψ(~ p3 , s3 ) and the ψ y with the ψ(~
p4 , s4 ) (see the left diagram
above). So, as per Remark 12.2.2, we want the ψ x closest to the bra-vector, which is how
our expression is already written. So for the t channel we don’t pick up any more minus
signs. However for the u channel we need flip the above contractions, that is contract ψ x with
ψ(~p4 , s4 ) and ψ y with ψ(~
p3 , s3 ). This means we need to place the ψ y closest to the bra-vector
and so we have to anticommute it past the ψ x . This is where our sign comes from!
For completeness, the full expressions for the t and u channel are
i(−ig)2 u(~
p3 , s3 )α u(~
p1 , s1 )α u(~
p4 , s4 )β u(~
p2 , s2 )β
(p1 − p3 )2 − µ2 + i
and
i(−ig)2 u(~
p4 , s4 )α u(~
p1 , s1 )α u(~
p3 , s3 )β u(~
p2 , s2 )β
− 2 2
(p1 − p4 ) − µ + i
respectively. So we don’t just swap p3 → p4 in the denominator, but we also include the
minus sign in front of the second expression for the Fermion statistics.
Remark 12.2.4 . Note that we have introduced labels α/β on the matrices so we can see what
is contracted with what, e.g. for the first expression the incoming ψ(~
p1 , s1 ) couples to the
outgoing ψ(~p3 , s3 ).
Remark 12.2.5 . Note it doesn’t matter which term we put the minus sign in front of, as
long as there is a relative minus sign between the two expressions. This is just because the
matrix element for a given process is given by a sum over all the different contributions (this
7
More accurately, we include a term x ↔ y in our expressions, so we fix the x and y and just calculate the
answer for this, then add the symmetry factors at the end.
8
Note we do this so that we cancel the minus factor in our expression above.
LECTURE 12. DIRAC PROPAGATORS & PERTURBATION THEORY 116
just comes from the fact that each diagrams corresponds to a different contraction type,
and we sum over the contractions). Now the two expressions above are just the individual
contributions to the matrix element, but the individual terms are not measurable; all we can
measure is the absolute value of the full matrix element squared. The difference between
Mt − Mu and Mu − Mt is clearly just an overall minus sign, and which is completely
redundant. That is
|Mt − Mu |2 = |Mu − Mt |2 .
12.3 QED
We end the course with a very brief discussion of QED. We do not have the time here to
discuss it in detail, but there is a separate QED course which will expand much further on
the theory. The aim here is basically just to introduce the Lagrangian and state the Feynman
rules.
Remark 12.3.1 . I have tried to expand on some of the equations/ideas presented in the course,
however there is still a significant amount of information that would be useful to know. As
always, basically all of this information can be found in Prof. Tong’s notes, chapter 6.
• The photon couples to electric charge, by which we mean the conserved charge for QED
is given by local conservation of electric charge. So things like electrons participate in
QED reactions, and, as the photon is electrically neutral, the only acceptable vertices for
an electron in QED is scattering (i.e. one electron in one electron out) or annihilation
(i.e. electron and positron in, photon out).
Remark 12.3.2 . Note that the fact that the photon is electrically neutral means that we
cannot have any photon-photon interactions. This is contrast to, say, φ4 theory where we had
a vertex with more than one φ. A better comparison to make would be to QCD (the QFT of
the strong force) where the gauge boson,10 called a gluon, is charged under the QCD charge
(called colour). This means that we can get gluon self interactions. This is not discussed at
all in this course, but just mentioned for interest.
9
Note physical photons only have 2 polarisations, so we should only have 2 degrees of freedom in Aµ ,
however we naïvely have 4 (one for each µ = 0, .., 3. It turns out you can remove two of these leaving only 2
degrees of freedom. We obtain this using gauge fixing, see Prof. Tong’s notes for more details.
10
Basically this just means the particle that describes the force. So the gauge boson for QED is the photon.
LECTURE 12. DIRAC PROPAGATORS & PERTURBATION THEORY 117
Ok so let’s couple the Maxwell theory to some Fermions. The Lagrangian is given by
putting together the Maxwell Lagrangian, Equation (12.8), and the Dirac Lagrangian, Equa-
tion (10.17), and then include a coupling between the fields. The result is
1
L = − Fµν F µν + ψ iD
/ − m)ψ Dµ := ∂µ + ieAµ . (12.9)
4
where we recall that D
/ := γ µ Dµ . We see that our coupling constant is −ieγ µ , and so we have
Z
HI = ie d3 ~x ψ Aψ
/ (12.10)
Remark 12.3.3 . The definition of Dµ in Equation (12.9) might seem a bit strange, however we
can show that this definition is required to preserve our internal symmetries. For an outline
of this, we set the following exercise.
Exercise
Show that Equation (12.9) is invariant under the transformation
Aµ → Aµ + ∂µ α, and ψ → e−ieα ψ
Dµ ψ → e−ieα Dµ ψ
p~
Outgoing Photon ε∗µ (p) = h~
p, ε| Aµ
Vertex −ieγ µ
~k −iηµν
Photon Propagatora
k 2 + i
a
In the Feynman gauge. We give a more general expression in the QED course.
We use these rules in the same way as for the previous theories to convert Feynman
diagrams into mathematical expressions. Note that we are considering couplings to Ferimions
(i.e. the Dirac fields) and so we have to include the factors of −1 between diagrams as per
the end of the last section.
Remark 12.3.4 . Note there is no mass term in the denominator of the propagator term. This
is because the photon is massless.
We end the course by giving one example of a Feynman diagram in QED and its corre-
sponding mathmatical expression.
Example 12.3.5 . As we stated in the bullet points above, photons couple to electric charge,
and so the scattering process e− e− → e− e− is valid. The only two valid diagrams are the t
channel and u channel, i.e.11
k and k
e− e− e− e−
−ieγ µ −ieγ µ p~4 , s4
p~2 , s2 p~4 , s4 p~2 , s2
11
The right diagram is a bit cramped, apologies. I just wanted to save time drawing a whole new diagram
so copied the previous one and made some small edits. I think its clear what labels what so decided not to
spend ages tweaking it.
LECTURE 12. DIRAC PROPAGATORS & PERTURBATION THEORY 119
(−iηµν )
p3 , s3 )α (−ieγ µ )u(~
u(~ p1 , s1 )α p4 , s4 )β (−ieγ µ )u(~
u(~ p2 , s2 )β
(p1 − p3 )2 + i
and
(−iηµν )
p4 , s4 )α (−ieγ µ )u(~
−u(~ p1 , s1 )α p3 , s3 )β (−ieγ µ )u(~
u(~ p2 , s2 )β
(p1 − p4 )2 + i
respectively. Note the minus sign in front of the second expression. This comes from the
exchange of the end state particles.
Exercise
Draw the Feynman diagrams and obtain the corresponding mathematical expression
for the QED process e− e+ → e− e+ . Hint: The answer to the last exercise might be
useful.
Exercise
There is something called a muon, which is roughly speaking just a heavier version
of the electron. We denote it by µ− . Importantly it is also electrically charged, and
so couples to the photon. Draw the Feynman diagrams for the scattering process
e− µ− → e− µ− , and write down their mathematical expressions. If you like you can
distinguish between electron and muon type particles by denoting electrons with u/vs
and muons with an a/bs.
Useful Texts & Further Readings
Textbooks
• M. E. Peskin and D. V. Schroeder, “An Introduction to QFT ” (Addison Wesley, 1995).
• Schwartz, Matthew D. Quantum field theory and the standard model (Cambridge Uni-
versity Press, 2014).
• Prof. Timo Weigand "Quantum Field Theory I + II", Institute for Theoretical Physics,
Heidelberg University.
120