We’ve built up the technology of field theory and path integrals, and I’ve promised you that this is sufficient to understand what happens at a second order phase transition. But so far, we’ve made little headway. All we’ve seen is that as we approach the critical point, fluctuations dominate and the Gaussian path integral is no longer a good starting point. We need to take the interactions into account.
Sometimes in physics, you can understand a phenomenon just by jumping in and doing the right calculation. And we will shortly do this, using perturbation theory to understand how the terms change the critical exponents. However, to really understand second order phase transitions requires something more: we will need to set up the right framework in which to think of physics at various length scales. This set of ideas was developed in the 1960s and 1970s, by people like Leo Kadanoff, Michael Fisher and, most importantly, Kenneth Wilson. It goes by the name of the Renormalisation Group.
Let’s start by painting the big picture. As in the previous section, we’re going to consider a class of theories based around a single scalar field in dimensions. (We will consider more general set-ups in Section 4.) The free energy takes the now familiar form,
(3.65) |
In what follows, we will look at what happens as we approach the critical point from above . All the important temperature dependence in (3.65) is sitting in the quadratic term, with
(3.66) |
In contrast to the previous section, we will allow to take either sign: in the disordered phase, and in the ordered phase where .
There is one important change in convention from our earlier discussion: we have rescaled the coefficient of the gradient term to be ; we will see the relevance of this shortly. All other terms have arbitrary coefficients.
So far we’ve focussed on just a few couplings, as shown in the free energy (3.65). Here we’re going to expand our horizons. We’ll consider all possible terms in the free energy, subject to a couple of restrictions. We’ll insist that the free energy is analytic around , so has a nice Taylor expansion, and we will insist on the symmetry , so that only even powers of arise. This means, for example, that we will include the term and and and and so so. Each of these terms comes with its own coupling constant. However, we don’t include terms like because this violates the symmetry, nor because this is not analytic at .
Next, consider the infinite dimensional space, parameterised by the infinite number of coupling constants. We will call this the theory space (although I should warn you that this isn’t standard terminology). You should have in your mind something like this:
But possibly bigger.
As we’ve seen, our interest is in computing the partition function
(3.67) |
Note that I’ve written the exponent as rather than . This is because the overall power of does nothing to affect the physics; all the relevant temperature dependence is in the coefficient (3.66) while, for the quantities of interest near the critical point, this overall factor can be set to . You can think that we’ve simply rescaled this into the field .
There is one more ingredient that we need to make sense of the path integral (3.67). This is the UV cut-off . Recall that, implicit in our construction of the theory is the requirement that the Fourier modes vanish for suitably high momenta
This arises because, ultimately our spins sit on some underlying lattice which, in turn, was coarse-grained into boxes of size . The UV cut-off is given by .
Until now, the UV cut-off has taken something of a back seat in our story, although it was needed to render some of the path integral calculation in the previous section finite. Now it’s time for to move centre stage. As we will explain, we can use the cut-off to define a flow in the space of theories.
Suppose that we only care about physics on long distance scales, . Then we have no real interest in the Fourier modes with . This suggests that we can write down a different theory, that has a lower cut-off,
for some . As long as , the scale of interest, our theory can tell us everything that we need to know. Moreover, we know, at least in principle, how to construct such a theory. We write our Fourier modes as
where describe the long-wavelength fluctuations
and describe the short-wavelength fluctuations that we don’t care about
There are several other names for these variables that are used interchangeably. The modes and are also referred to as low- and high-energy modes or, importing language from quantum mechanics, slow and fast modes, respectively. In a rather quaint nod to the electromagnetic spectrum, the short-distance, microscopic physics that we care little about is often called the ultra-violet; the long-distance physics that we would like to understand is the infra-red.
Similarly, we decompose the free energy, written in Fourier space, as
Here involves the terms which mix the short and long-wavelength modes. The partition function (3.67) can then be written as
We write this as
where is known as the Wilsonian effective free energy. (In fairness, this term is rarely used: you’re more likely to hear “Wilsonian effective action” to describe the analogous object in a path integral describing a quantum field theory.) We’re left with a free energy which describes the long-wavelength modes, but takes into account the effects of the short wavelength modes. It is defined by
(3.68) |
In subsequent sections, we’ll put some effort into calculating this object. However, at the end of the day the new free energy must take the same functional form as the original free energy (3.65), simply because we started from the most general form possible. The only effect of integrating out the high-momentum modes is to shift the infinite number of coupling constants, so we now have
(3.69) |
We would like to compare the new free energy (3.69) with the original (3.65). However, we’re not quite there yet because, the two theories are different types of objects – like apples and oranges – and shouldn’t be directly compared. This is because the theory is defined by both the free energy and the UV cut-off and, by construction our two theories have different cut-offs. This means that the original theory can describe things that the new theory cannot, namely momentum modes above the cut-off .
It is straightforward to remedy this. We can place the two theories on a level playing field by rescaling the momenta in the new theory. We define
Now takes values up to , as did in the original theory. The counterpart of this scaling in real space is
This means that all lengths scales are getting smaller. You can think of this step as zooming out, to observe the system on larger and larger length scales. As you do so, all features become smaller.
There is one final step that we should take. The new theory will typically have some coefficient in front of the leading, quadratic gradient term. To compare with the original free energy (3.65), we should rescale our field. We define
Which, in position space, reads
(3.70) |
Now, finally, our free energy takes the form
(3.71) |
We see that this procedure induces a continuous map from onto the space of coupling constants. Our original coupling constants in (3.65) are those evaluated at . As we increase , we trace out curves in our theory space, that look something like the picture shown in Figure 22
We say that the coupling constants flow, where the direction of the flow is telling us what the couplings look like on longer and longer length scales. The equations which describe these flows – which we will derive shortly – are known, for historic reasons, as beta functions.
These, then are the three steps of what is known as the renormalisation group (RG):
Integrate out high momentum modes, .
Rescale the momenta .
Rescale the fields so that the gradient term remains canonically normalised.
You may wonder why we didn’t just include a coupling constant for the gradient term, and watch that change too. The reason is that we can always scale this away by redefining . But is just a dummy variable which is integrated over the path integral, so this rescaling can’t change the physics. To remove this ambiguity, we should pin down the value of one of the coupling constants, and the gradient term is the most convenient choice. If we ever find ourselves in a situation where for some then we would have to re-evaluate this choice. (We’ll actually come across an example where it’s sensible to make a different choice in Section 4.3.)
The “renormalisation group” is not a great name. It has a hint of a group structure, because a scaling by followed by a scaling by gives the same result as a scaling by . However, unlike for groups, there is no inverse: we can only integrate out fields, we can’t put them back in. A more accurate name would be the “renormalisation semi-group”.
The procedure we’ve described above is the renormalisation group in momentum space: to get an increasingly coarse-grained description of the physics, we integrate out successive momentum shells. This version of the renormalisation group is most useful when dealing with continuous fields and will be the approach we will focus on in this course.
There is a somewhat different, although ultimately equivalent, phrasing of the renormalisation group which works directly in real space. This approach works best when dealing directly with lattice systems, like the Ising model. As we explained rather briefly in Section 1.3, one constructs a magnetisation field by coarse-graining over boxes of size , each of which contains many lattice sites. One can ask how the free energy changes as we increase , a procedure known as blocking. Ultimately this leads to the same picture that we built above.
Even before we do any calculations, there are general lessons to be extracted from the framework above. Let’s suppose we start from some point in theory space. This can be arbitrarily complicated, reflecting the fact that it contains informations about all the microscopic, short-distance degrees of freedom.
Of course, we care little about most of these details so, in an attempt to simplify our lives we perform a renormalisation group transformation, integrating out short distance degrees of freedom to generate a new theory which describes the long wavelength physics. And then we do this again. And then we do this again. Where do we end up?
There are essentially two possibilities: we could flow off to infinity in theory space, or we could converge towards a fixed point. These are points which are invariant under a renormalisation group transformation. (One could also envisage further possibilities, such as converging towards a limit cycle. It turns out that these can be ruled out in many theories of interest.)
Our interest here lies in the fixed points. The second step in the renormalisation group procedure ensures that fixed points describe theories that have no characteristic scale. If the original theory had a correlation length scale , then the renormalised theory has a length scale . (We will derive this statement explicitly below when we stop talking and start calculating.) Fixed points must therefore have either or .
In the disordered phase, with , enacting an RG flow reduces the correlation length. Pictorially, we have
In this case, shrinking the correlation length is equivalent to increasing the temperature. The end point of the RG flow, at , is the infinite temperature limit of the theory. This is rather like flowing off to infinity in theory space. As we will see, it is not uncommon to end up here after an RG flow. But it is boring.
Similarly, in the ordered phase the RG flow again reduces the correlation length,
Now the end point at corresponds to the zero temperature limit; again, it is a typical end point of RG flow but is dull.
Theories with are more interesting. As we saw above, this situation occurs at a critical point where the theory contains fluctuations on all length scales. Now, if we do an RG flow, the theory remains invariant. In terms of our visual configurations,
Note that the configuration itself doesn’t stay the same. (It is, after all, merely a representative configuration in the ensemble.) However, as the fluctuations on small distance scales shrink away due to RG, they are replaced by fluctuations coming in from larger distance scales. The result is a theory which is scale invariant. For this reason, the term “critical point” is often used as a synonym for “fixed point” of the RG flow.
This picture is all we need to understand the remarkable phenomenon of universality: it arises because many points in theory space flow to the same fixed point. Thus, many different microscopic theories have the same long distance behaviour.
It is useful to characterise the properties of fixed points by thinking about the theories in their immediate neighbourhood. Obviously, there are an infinite number of ways we can move away from the fixed point. If we move in some of these directions, the RG flow will take us back towards the fixed point. These deformations are called irrelevant because if we add any such terms to the free energy we will end up describing the same long-distance physics.
In contrast, there will be some directions in which the RG flow will sweep us away from the fixed point. These deformations are called relevant because if we add any such terms to the free energy, the long-distance physics will be something rather different. Examples of relevant and irrelevant deformations are shown in Figure 23. Much of the power of universality comes from the realisation that the vast majority of directions are irrelevant. For a given fixed point, there are typically only a handful of relevant deformations, and an infinite number of irrelevant ones. This means that our fixed points have a large basin of attraction, huge slices of the infinite dimensional theory space all converging to the same fixed point. The basin of attraction for a particular fixed point is called the critical surface.
Finally, it’s possible that our fixed point is not a point at all, but a line or a higher dimensional surface living within theory space. In this case, if we deform the theory in the direction of the line, we will not flow anywhere, but simply end up on another fixed point. Such deformations are called marginal; they are rare, but not unheard of.
Universality is a wonderful thing if you want to understand the low-energy, long-wavelength physics. It tells you that you can throw away many of the microscopic details because they are irrelevant for the things that you care about.
In contrast, if you want to understand the high-energy, short distance physics then universality is the devil. It tells you that you have very little hope of extracting any information about microscopic degrees of freedom if you only have access to information at long distances. This is because many different microscopic theories will all give the same answer.
As we saw in Section 2.3, quantum field theory is governed by the same mathematical structure as statistical field theory, and the comments above also apply. Suppose, for example, that you find yourself living in a technologically adolescent civilisation that can perform experiments at distance scales of cm or so, but no smaller. Yet, what you really care about is physics at, say, cm where you suspect that something interesting is going on. The renormalisation group says that you shouldn’t pin your hopes on learning anything from experiment.
The renormalisation group isn’t alone in hiding high-energy physics from us. In gravity, cosmic censorship ensures that any high curvature regions are hidden behind horizons of black holes while, in the early universe, inflation washes away any trace of what took place before. Anyone would think there’s some kind of conspiracy going on….
The idea that second order phase transitions coincide with fixed points of the renormalisation group is a powerful one. In particular, it provides an organising principle behind the flurry of critical exponents that we met in Section 1.
As we explained above, at a fixed point of the renormalisation group any scale must be washed away. This is already enough to ensure that correlation functions must take the form of a power-law,
(3.72) |
Any other function would require a scale on dimensional grounds. The only freedom that we have is in the choice of exponent which we have chosen to parameterise as . One of the tasks of the RG procedure is to compute , and we will see how this works in Section 3.5.
However, even here there’s something of a mystery because usually we can figure out the way things scale by doing some simple dimensional analysis. (If you would like to refresh your memory, some examples of dimensional analysis can be found in Chapter 3 of the lectures on Dynamics and Relativity.) What does that tell us in the present case?
We will measure dimension in units of inverse length. So, for example, while . The quantity must be dimensionless because it sits in the exponent of the partition function as . The first term is
From this we learn that
(3.73) |
Which, in turn, tells us exactly what the exponent of the correlation function must be: .
This is sobering. Dimensional analysis is one of the most basic tools that we have, and yet it seems to be failing at critical points where experiment is showing that . What’s going on?
A better way to think about dimensional analysis is to think in terms of scaling. Suppose that we rescale all length as . How should other quantities scale so that formula remains invariant? Stated this way, it’s clear that there’s a close connection between dimensional analysis and RG. The correlation function (3.72) is telling us that we should rescale , where
(3.74) |
This is called the scaling dimension. It differs from the naive “engineering dimension” by the extra term which is referred to as the anomalous dimension.
We still haven’t explained why the scaling dimension differs from engineering dimension. The culprit turns out to be the third step of the RG procedure (3.70) where the field gets rescaled. In real space, this is viewed as coarse-graining over blocks of larger and larger size . As we do so, it dresses with this UV cut-off scale , often in a complicated and non-intuitive way. This means that the correlation function (3.72) is actually
which is in full agreement with naive dimensional analysis. We can work with usual engineering dimensions if we keep track of this microscopic distance scale . But it is much more useful to absorb this into and think of a coarse-grained observable, with dimension , that is the appropriate for measuring long distance correlations.
The critical exponents that we met in Section 1.2.3 are all a consequence of scale invariance, and dimensional analysis based on the scaling dimension. Let’s see how this arises.
We know that as we move away from the critical point by turning on , we introduce a new length scale into the problem. This is the correlation length, given by
(3.75) |
Here is called the reduced temperature, while is another critical exponent that we will ultimately have to calculate. Since is a length scale, it transforms simply as . In other words, it has scaling dimension . The meaning of the critical exponent is that the reduced temperature scales as , with
(3.76) |
In what follows, our only assumption is that the correlation length is the only length scale that plays any role.
We start with the thermodynamic free energy, , evaluated at . This takes the form
Because is scale invariant at the fixed point, must have scaling dimension , which immediately tells us that
There is an intuitive way to understand this. At close to , the spins are correlated over distances scales , and can be viewed as moving as one coherent block. The free energy is extensive, and so naturally scales as .
From the thermodynamic free energy, we can compute the singular contribution to the heat capacity near . It is
where the second relationship is there to remind us that we already had a name for the critical exponent related to heat capacity. We learn that
(3.77) |
This is called the Josephson relation or, alternatively, the hyperscaling relation.
The next critical exponent on the list is . Recall that this relates the magnetisation in the ordered phase – which we used to call and have now called – to the temperature as
But the scaling dimensions of this equation only work if we have
(3.78) |
The next two critical exponents require us to move away from the critical point by turning on a magnetic field . This is achieved through the addition of a linear term in the free energy. (We didn’t include such a linear term in our previous discussion of RG, but it can be added without changing the essence of the story.) The scaling dimensions of this term must add to zero, giving
Now we can look at the various relationships. The behaviour of the susceptibility near the critical point is
Once again, the scaling dimensions are enough to fix to be
(3.79) |
which is sometimes called Fisher’s identity. Once again, there is an intuitive way to understand this. The meaning of is that the spins are no longer correlated at distances . This can be seen, for example, in our original formula (2.62). Using our earlier expression (2.61) for the susceptibility, we have
which again gives .
The final critical exponent relates the magnetisation to the magnetic field when we sit at the critical temperature . It should come as little surprise by now to learn that this is again fixed by scaling analysis
(3.80) |
We end up with four equations, relating (3.77), (3.78), (3.79) and (3.80) to the critical exponents and . For convenience, let’s recall what values we claimed this exponents take:
MF | 1 | 3 | 0 | |||
---|---|---|---|---|---|---|
0 | 15 | 1 | ||||
0.1101 | 0.3264 | 1.2371 | 4.7898 | 0.0363 | 0.6300 |
where we’ve used the result (2.47), including quadratic fluctuations, for the mean field value of . We see that the relations are satisfied exactly for and to within the accuracy stated for . However, there’s a wrinkle because they only agree with the mean field values when !
This latter point is an annoying subtlety and will be explained in Section 3.3.2. Our main task is to understand why the mean field values don’t agree with experiment when .
The kind of dimensional analysis above also determines whether a given interaction is relevant, irrelevant or marginal.
Consider an interaction term in the free energy,
(3.81) |
Here can be or or any of the other infinite possibilities. In a spillover from quantum field theory, the different interaction terms are referred to as operators.
We’re interested in operators which, in the vicinity of a given point, transform simply under RG. Specifically, suppose that, under the rescaling , the operator has a well defined scaling dimension, transforming as
(3.82) |
You can think of such operators as eigenstates of the RG process. From the free energy (3.81), the scaling dimension of the coupling is
Under an RG flow, these couplings scale as . We can see immediately that either diverges or vanishes as we push forwards with the RG. Invoking our previous classification, is:
Relevant if
Irrelevant if
Marginal if
The tricky part of the story is that it’s not always easy to identify the operators which have the nice scaling property (3.82). As we’ll see in the examples below, these are typically complicated linear combinations of the operators and and so on.
It’s now time to start calculating. We will start by sitting at a special point in theory space and enacting the renormalisation group. At this special point, only two quadratic terms are turned on:
(3.83) |
where we’ve added a subscript to the coefficient in anticipation the fact that this quantity will subsequently change under RG flow.
Because the free energy is quadratic in , it has the property that there is no mixing between the short and long wavelength modes, and so factorises as
Integrating over the short wavelength modes is now easy, and results in an overall constant in the partition function
This constant doesn’t change any physics; it just drops out when we differentiate to compute correlation functions. However, we’re not yet done with the our RG; we still need to do the rescaling
(3.84) |
where is constant that we will determine. Written in terms of the rescaled momenta, we have
We can put this back in the form we started with if we take
(3.85) |
leaving us with
The only price that we’ve paid for this is that the coefficient of quadratic term has become
(3.86) |
This illustrates how the length scales in the problem transform under RG. Recall that the correlation length (2.55) is . We see that, under an RG procedure,
The fixed points obey . As we anticipated previously, there are two of them. The first is . This corresponds to a state which has infinite temperature. It is not where our interest lies. The other fixed point is at . This is known as the Gaussian fixed point.
As we mentioned previously, we would like to classify fixed points by thinking about what happens when you sit near them. Do you flow into the fixed point, or get pushed away?
We already have the answer to this question in one direction in coupling space. If we add the term , the scaling (3.86) tells us that gets bigger as we flow towards the infra-red. This is an example of a relevant coupling: turning it on pushes us away from the fixed point.
Here is another example: it is simple to repeat the steps above including the term in the free energy. Upon RG, this coupling flows as . It is an example of an irrelevant coupling, one which becomes less important as we flow towards the infra-red.
More interesting are the slew of possible couplings of the form
(3.87) |
where, to keep the symmetry, we restrict the sum to even. Here things are a little more subtle because, once we turn these couplings, on the first step of the RG procedure is no longer so simple. Integrating out the short distance modes will shift each of these couplings,
We will learn how to calculate the in section 3.4. But, for now, let’s ignore this effect and concentrate on the second and third parts of the RG procedure, in which we rescale lengths and fields as in (3.84). In this approximation, the operators enjoy the nice scaling property (3.82),
The free energy is then rescaled by
To restore the coefficient of the gradient term, we pick the scaling dimension
For once, the scaling dimension coincides with the engineering dimension (3.73): . This is because we’re looking at a particularly simple fixed point. Note that this is related to our earlier result (3.85) by , with the extra factor of coming from the in the definition of the Fourier transform.
Our free energy now takes the same form as before,
where
(3.88) |
We see that the way these coupling scale depends on the dimension . For example, the coefficient for scales as
We learn that is irrelevant for and is relevant for . According to the analysis above, when , we have and the coupling is marginal. In this case, however, we need to work a little harder because the leading contribution to the scaling will come from the corrections that we neglected. We’ll look at this in the next section.
Restricting to the plane of couplings parameterised by and , we see that (if we neglect the interactions) the RG flow near the origin is very different when and . These are shown in the two figures. In the former case, we need to tune only if we want to hit the fixed point; the other couplings will take care of themselves. In contrast, when both of these couplings are relevant. This means that we would need to tune both to zero if we want to hit the Gaussian fixed point.
We can tally this with our discussion in Section 3.2. The fact that the scaling dimension coincides with the naive engineering dimension immediately tells us that . Meanwhile, the scaling of is given by , which tells us that . From this we can use (3.77) - (3.80) to extract the remaining critical exponents. These agree with mean field for , but not for . (We will address the situation in in Section 3.3.2.)
It is no coincidence that this behaviour switches at , which we previously identified as the upper critical dimension. In an experiment, one can always change by varying the temperature. However, one may not have control over the couplings which typically correspond to some complicated microscopic property of the system. If is irrelevant, we don’t care: the system will drive itself to the Gaussian fixed point. In contrast, if is relevant the system will drive itself elsewhere. This is why we don’t measure mean field values for the critical exponents: these are the critical exponents of the Gaussian fixed point.
The coupling for scales as
This is irrelevant in , relevant in and, naively, marginal in .
Note that in dimension all of the couplings are relevant.
So far, this all looks rather trivial. However, things become much more interesting at other fixed points. For example, around most fixed points . Indeed, around most fixed points neither nor will have well defined scaling dimension; instead those operators to which one can assign a scaling dimension consist of some complicated linear combination of the . We will start to understand this better in Section 3.4.
The meaning of the phrase “mean field theory” has evolved as these lectures have progressed. We started in Section 1.1.2 by introducing mean field as a somewhat dodgy approximation to the partition function. Subsequently, we used the expression “mean field theory” to mean writing down a free energy and focussing on the saddle point equations. This saddle point is a good approximation to the partition function only when the couplings are small; this is true only in the vicinity of the Gaussian fixed point. For this reason, using mean field theory is usually synonymous with working at the Gaussian fixed point, and ignoring the effect of operators like on fluctuations. (Of course, mean field still retains the term in the ordered phase, where it is needed to stabilise the potential.)
Until now, we have have restricted ourselves to interactions with even, to zealously safeguard the symmetry . One particularly nice aspect of RG is that if we restrict ourselves to a class of free energies that obey a certain symmetry, then we will remain in that class under RG. We’ll see examples of this in Section 3.4.
However, suppose that we sit outside of this class and turn on interactions with odd. The leading order effect is the magnetic field that we included in our original Ising model. This is always a relevant interaction. This means that if we want to hit the critical point, we must tune this to zero.
It may be more natural to tune in some systems that others. For example, a magnet in the Ising class automatically has unless you choose to submit it to a background magnetic field. This means that it’s easy to hit the critical point: just heat up a magnetic and it will exhibit a second order phase transition.
In contrast, in the liquid gas system, setting “” is less natural. Unlike in the Ising model, there is no symmetry manifest in the microscopic physics of gases. Instead, it is an emergent symmetry which relates the density of liquid and gas states at the phase transition. Correspondingly, if we simply take a liquid and heat it up then we’re most likely to encounter a first order transition, or no transition at all. If we want to hit the critical point, we must now tune the two relevant operators: temperature and pressure, which corresponds to the linear term with coefficient .
In both situations above we really need to tune two relevant couplings to zero to hit the critical point. Of these, one is even under and one is odd under . Doing this will allow us to hit a fixed point with two relevant deformations, one even one odd. This is the Gaussian fixed point in and is something else (to be described below) in .
What about higher order interactions with odd. If we have to tune , do we not also need to tune ? It turns out that that the interaction is redundant. If you have a free energy with no symmetry, and all powers of , then you can always redefine your field as for some constant . This freedom allows you to eliminate the term. Note that if your free energy enjoys the symmetry then it prohibits you from making this shift.
We’ve learned that the interaction is irrelevant for , and so one can hit the Gaussian fixed point by tuning just one parameter: .
However, there’s one tricky issue that we haven’t yet explained: the mean field exponents agree with the scaling analysis of Section 3.2 only when . Comparing the two results, we have
MF | 1 | 3 | 0 | |||
---|---|---|---|---|---|---|
Scaling | 1 | 0 |
where we’ve used the result (2.47), including quadratic fluctuations, for the mean field value of . This agrees with the scaling analysis. However, for , the exponents and differ. It turns out that the results from Landau mean field are correct, and those from the scaling analysis are wrong. Why?
To understand this, let’s recall our scaling argument from Section 3.2. We set and focus on the critical exponent . The magnetisation scales with the temperature as
Here is identified with the scalar field . Scaling analysis gives . But both mean field and scaling analysis agree that and , and this gives , rather than the mean field result .
However, we were a little quick in the scaling analysis because we neglected the quartic coupling . Mean field really told us (1.31),
But both and scale under RG flow. The scaling dimension of is and now the mean field result, with is fully compatible with scaling.
There’s a more general lesson to take from this. It is tempting, when doing RG, to think that we can just neglect the irrelevant operators because their coefficients flow to zero as we approach the infra-red. However, sometimes we will be interesting in quantities – such as the magnetisation above – which have the irrelevant coupling constants sitting in the denominator. In this case, one cannot just blindly ignore these irrelevant couplings as they affect the scaling analysis. When this happens, the irrelevant coupling is referred to as dangerously irrelevant.
This is a good point to revisit an issue that we previously swept under the rug. We started our discussion with a lattice model, but very quickly moved to the continuum, field theory. Along the way we stated, without proof, that we expect the long distance physics to enjoy rotational invariance and we restricted our attention to field theories with this property. Why are we allowed to do this?
To make the discussion concrete, consider a square lattice in dimensions. This has a discrete rotational symmetry, together with a reflection symmetry. These combine together into the dihedral group .
Our field theory description will respect the symmetry of the underlying lattice model, together with the symmetry which ensures that fields come in pairs. But this would appear to be much less powerful than the full continuous rotation and reflection symmetry. Have we cheated?
Let’s see what kind of terms we might expect. First, there are some simple terms that are prohibited by the dihedral symmetry. For example a lone term would break the discrete rotational symmetry and so would not appear in the free energy. Similarly, a term breaks the symmetry. (On top of this, it is also a total derivative and so doesn’t contribute to the free energy.) The lowest dimension term that includes derivatives and is compatible with the discrete symmetry is
But this term happens to be invariant under the full, continuous rotational symmetry. We should keep going. The first term that preserves , but not , is
There is no reason not to add such terms to the free energy and, in general, we expect that these will be present in any field theoretic description that accurately describes the microscopic physics. However, this operator has dimension and so is irrelevant. This means that it gets washed away by the renormalisation group, and the long wavelength physics is invariant under the full symmetry. We say that the continuous rotational symmetry is emergent. A similar argument holds for higher dimensions.
The previous section left two questions hanging. What happens to the renormalisation of the coupling in dimensions? And where does the flow of take us in dimensions? In this section we will answer the first of these. In Section 3.5 we will see that our analysis also contains the answer to the second.
We now repeat our RG procedure, but with a different starting point in theory space,
The renormalisation group procedure tells us to split the Fourier modes of the field into long and short wavelengths,
(3.89) |
and write the free energy as
where we take to coincide with the quadratic terms (3.83), and the interaction terms are
Note that we’ve chosen to include, for example, in the interaction terms rather than . This is a matter of convention.
The effective free energy for , defined in (3.68), is given by
There is a nice interpretation of this functional integral . We can think of it as computing the expectation value of , treating as the random variable with Gaussian distribution . In other words, we can write this as
where the subscript on is there to remind us that we are averaging over the modes only. We take the definition of the path integral to be suitably normalised so that . Taking the log of both sides,
(3.90) |
Our task is to compute this.
We can’t do this functional integral exactly. Instead, we resort to perturbation theory. We assume that is suitably small, and expand. (The dimensionless small parameter is .) We first Taylor expand the exponential,
and we then Taylor expand , to get
(3.91) |
where, in general, the term is cumulant of . This also follows from the same kind of manipulations that we did at the beginning of Section 2.2. We will deal with each of terms above in turn.
At leading order in , we need to compute . The first order of business is to expand the interaction terms (3.90) in Fourier modes. We have
There are five different “terms with ”, most of which do not give anything interesting. These five terms are:
: This term doesn’t include any , the average is trivial. It carries over to give the term in the effective free energy.
: This term has just a single and so vanishes when averaged over the Gaussian ensemble.
: This term is interesting. We will look at it more closely below. For now, note that the factor of 6 comes from the different choices of momentum labels.
: This term is cubic in and, like any term with an odd number of insertions, vanishes when averaged over the Gaussian ensemble.
: This term doesn’t include any , it simply gives constant to the free energy. It will not be important here.
We learn that we need to compute just a single term,
(3.92) |
But this is the same kind of correlation function that we computed in Section 2.2: it is given by
(3.93) |
After playing around with the delta-functions, and relabelling momentum variables, we’re left with our first correction to the free energy,
where the limits on the integral reflect the fact that we’ve only integrated out the short wavelength modes, whose momenta lie within this band. We see that, at order , we get a correction only to the quadratic term whose coefficient becomes
(3.94) |
Finally, we should enact the rescaling steps of the renormalisation group. This takes the same form as before (3.84),
This gives the same scaling of the parameters that we saw in Section 3.3. We have
(3.95) |
The upshot of this calculation is that turning on a coupling will give rise to a quadratic coupling under RG flow. This is typical of these kinds of calculations: couplings of one type will induce others.
The coefficient of the term is particularly important for our story, since the critical point is defined to be the place where this vanishes. We see that it’s not so easy to make this happen. You can’t simply set at some high scale and expect to hit criticality. Indeed, the result (3.95) tells us that, at long wavelengths, the “natural” value is , which is typically large. If you want to hit the critical point, you must “fine tune” the original coefficient to cancel the new terms that are generated by RG flow.
You might think that this calculation answers the question of what happens to the theory in when we turn on . It certainly tells us that turning on this coupling will induce the relevant coupling and so take us away from the Gaussian fixed point. However, a closer look at (3.95) reveals that it’s possible to turn on a combination of and , so that remains zero. This combination is a marginal coupling. We learn that, at this order, there remains one relevant and one marginal deformation.
The corrections to the terms first arise at order . Here we have the contribution
(3.96) |
Expanding out , we find 256 different terms. We will see how to organise them shortly, but for now we make a few comments before focussing on the term of interest.
Some of the terms in will result in corrections that cannot be written as a local free energy, but are instead of the form for some . These terms will be cancelled by the terms. This is a general phenomena which you can learn more about in the lectures on Quantum Field Theory. In terms of Feynman diagrams, which we will introduce below, these kind of terms correspond to disconnected diagrams.
The terms that we care about in are those which can be written as local corrections to the free energy. Of immediate interest for us are a subset of terms in , given by
Let’s explain what’s going on here. Each is decomposed into Fourier modes and . The same is true for each . In the above term, we have chosen two out of the and two out of the ; the remaining terms are . Each combinatoric factor out front reflects the choice of picking two from . Meanwhile, the two delta functions come from doing the and integrals respectively. Matching the momenta in the arguments of the delta functions to the tells us that we’ve picked two from the and two from (as opposed to, say, all four from ).
To proceed, we need to compute the four-point function . To do this we need a result known as Wick’s theorem.
As we proceed in our perturbative expansion, the integrals start to blossom. From the form of the expansion (3.91), we can see that the integrand will involve expectation values of the form . There is a simple way to compute expectation values of this type in Gaussian ensembles. This follows from:
Lemma: Consider variables drawn from a Gaussian ensemble. This means that, for any function , the expectation value is
for some invertible matrix . The normalisation factor is and ensures that . The following identity then holds:
(3.98) |
for any constant .
Proof: This is straightforward to show since we can just evaluate both sides
where, in the last step, we used the fact that .
The Taylor expansion of the identity (3.98) gives us the expressions that we want. The left-hand-side is
Meanwhile, the right-hand-side is
Now we just match powers of on both sides. We immediately learn that
Our real interest is in even. Here we have to be a little careful because multiplying by the string of ’s automatically symmetrises the products of over the indices. So, for example, comparing the terms gives
(3.99) |
It’s not hard to convince yourself that
This leaves us with a sum over all pairwise contractions. This result is known as Wick’s theorem.
We can now apply Wick’s theorem to our free energy (3.4.2),
Recall that each of these propagators comes with a delta function,
The trick is to see how these new delta functions combine with the original delta functions in (3.4.2). There are two different structures that emerge. The first term, gives (ignoring factors of for now)
(3.100) | |||||
We’re still left with two delta functions over the variables. This means that when we go back to real space, this term does not become a local integral. Instead, if you follow it through, it becomes a double integral of the form . As we explained after (3.96), these terms are ultimately cancelled by corresponding terms in . They are not what interests us.
Instead, we care about the second and third terms in the Wick expansion of . Each of them gives a contribution of the form
(3.101) | |||||
where, in going to the last line, we have done the integral and relabelled . Now we have just a single delta function over and, correspondingly, when we go back to real space this will give a local contribution to the free energy. Indeed, the terms (3.4.2) now become
(3.102) |
where the factor of in (3.4.2) has disappeared because we get two contributions from the Wick expansion, each of which gives the same contribution (3.101). The remaining integral over is hidden in the function , which is given by
(3.103) |
This is not as complicated as it looks. We can write it as
All the terms that depend on the external momenta and will generate terms in the free energy of the form . These are irrelevant terms that will not be interesting for us other than to note that once we let loose the dogs of RG, we will no longer sit comfortably within some finite dimensional subspace of the coupling constants. Integrating out degrees of freedom generates all possible terms consistent with symmetries; flowing to the IR allows us to focus on the handful of relevant ones.
The contribution (3.102) to the free energy is what we want. Translating back to real space, we learn that the quadratic term gets corrections at this order. We have
(3.104) |
The minus sign in (3.104) is important. It can be traced to the minus sign in (3.96), and it determines the fate of the would-be marginal coupling in dimensions. Recall that, in , there is no contribution to the running of from the second and third steps of RG. Here, the leading contribution comes from the first step and, as we see above, this causes to get smaller as increases. This means that the theory in is similar in spirit to those in , with an irrelevant coupling.
However, in , the RG flow for happens much more slowly than other couplings. For this reason it is sometimes called marginally irrelevant, to highlight the fact that it only failed to be marginal when the perturbative corrections were taken into account. This is a general phenomenon: most couplings which naively appear marginal will end up becoming either marginally relevant or marginally irrelevant due to such corrections. In the vast majority of cases, the coupling turns out to be marginally irrelevant. However, there are a number of very important examples – the Kondo effect and non-Abelian gauge theories prominent among them – where a marginal coupling turns relevant. We’ll see such an example in Section 4.3.
Finally, just because is marginally irrelevant in does not mean that you can turn it on and expect to flow back to the Gaussian fixed point. As depicted in the diagram, the coupling mixes with . If you want to flow back to the Gaussian fixed point, you need to turn on a particular combination of and .
The calculation above needed some care. As we go to higher order terms in the expansion, the number of possibilities starts to blossom. Fortunately, there is a simple graphical formalism to keep track of what’s going on.
Suppose that we’re interested in a term in the expansion (3.91) of the form . This can be represented by one or more Feynman diagrams. Here is the game:
Each is represented by an external, solid line.
Each is represented by a dotted line.
The dotted lines are connected with each other to form internal loops. They are paired up in all possible ways, reflecting the pair contraction of Wick’s theorem. No dotted lines can be left hanging which means that the diagrams only make sense for even.
Each factor of is represented as a vertex at which four lines meet.
Each line has an attached momentum which is conserved as we move around the diagram.
Each of these diagrams is really shorthand for an integral. The dictionary is as follows:
Each internal line corresponds to an insertion of the propagator defined in (3.93)
For each internal loop, there is an integral .
Each vertex comes with a power of where the delta function imposes momentum conservation, with the convention that all momenta are incoming.
There are a bunch of numerical coefficients known as symmetry factors.
This means that a term in the effective action of the form will correspond to a diagram with external lines and vertices. We can see what these diagrams look like for some of the terms that we’ve met so far. At order , the rules don’t allow us to draw diagrams with an odd number of . The term in the expansion gives the trivial contribution to . In diagrams, it is
(3.105) |
Similarly, the terms are , but because these are internal lines we should join them up to get a diagram that looks like .
The interesting term at order is , which includes both and . This was evaluated in (3.92). It is represented by the diagram
where the integral arises because the momentum of the excitation running in the loop is not determined by momentum conservation of the external legs. Such corrections are said to arise from loop diagrams, as opposed to the tree diagrams of (3.105).
Finally, at order , the correction (3.102) to the coupling comes from the
Any graph that we can draw will appear somewhere in the expansion of given in (3.91). As we noted above, this is a cumulant expansion which has the rather nice graphical interpretation that only connected graphs appear in the expansion. For example, at order , the disconnected graph that appears in the expansion of that looks like will be cancelled by the same disconnected graph appearing in .
In the language of quantum field theory, it’s tempting to view the lines in the Feynman diagrams as the worldlines of particles. There is no such interpretation in the present case: they’re simply useful.
We can now also look at other diagrams and see what role they play. For example, you might be worried about the diagram . This is strictly vanishing, because the incoming momentum for lone leg is forced to be equal to the intermediate momentum for the propagator. Yet the momentum for and can never be equal .
There are, however, two further diagrams that we neglected which do have an interesting role to play. Both of these are two loop diagrams. They will not affect what we’re going to do later but, nonetheless, are worth highlighting. The first diagram is
for some whose exact form will not be important. This gives rise to a shift in the quadratic term, so that (3.94) is replaced by
The second diagram is
(3.106) |
We’ve called the result of this diagram ; again we won’t need its detailed form. Importantly, it is now a function of the external momentum . This means that it gives rise to two effects that are (in a technical sense) relevant. The first is that there is yet another renormalisation of , this time depending on . The second is novel: upon Taylor expanding , we get a contribution to the the gradient term, which is now
This, in turn, means that we need a new rescaling of the field. To order , we can write this as
This last, additional step is known as field renormalisation. (Actually, that’s not completely true. It should be known as “field renormalisation”, but instead is known as “wavefunction renormalisation”. This a terrible name, one that betrays the long and deep confusion that permeated the origins of this subject. Even in the context of quantum field theory, this rescaling has nothing to do with wavefunctions. It is a rescaling of fields!)
Although we won’t compute this field renormalisation exactly, it is nonetheless important for this is what gives rise to the anomalous dimension of , and this was underlying the whole scaling analysis of Section 3.2.
It is useful to write down equations which describe the flow of the coupling constants. These are first order differential equations which, for historic reasons are known as beta functions. It turns out to be convenient to parameterise the change in cut-off as
The renormalisation group transformation described above tells us that each coupling changes with scale, . The beta function is defined as
Note that increases as we flow towards the IR. This means that a positive beta function tells us that gets stronger in the IR, while a negative beta function means that gets weaker in the IR. (As an aside: this is the opposite to how beta functions are sometimes defined in quantum field theory, where one parameterises the flow in terms of energy rather than length.)
Before we jump straight in, it’s useful to take a step backwards and build up the beta functions. Let’s go back to our original scaling analysis around the Gaussian fixed point (3.88), where the running of the couplings is given by . The beta functions are
(3.107) |
Notice that, at this leading order, there’s no mixing between different couplings: turning on one coupling does not induce another to flow. As we saw above, this state of affairs no longer holds when we include interactions.
We now focus on the two most important couplings, and . At order , the RG equations are given by (3.95); the additional correction at order , given in (3.104), means that these get replaced by
(3.108) |
where
Note that we have kept our original scaling dimensions in (3.108); the corrections in scaling due to the diagram (3.106) will be subleading and not needed in what follows.
When we differentiate and to derive the beta functions, we will get two terms: the first is (3.107) and comes from the scaling; the second comes from the corrections, given by the integrals and . Differentiating these integrals is particularly easy. For small , we we write
Let’s restrict to dimensions. The beta function equations are
(3.109) |
These don’t (yet) contain any new physics, but it’s worth reiterating what information we can extract from these equations.
First, the beta function for has two terms; the first term comes from the second and third steps in RG (scaling), while the second comes from the first step in RG (integrating out). Meanwhile, the beta function for has only a single term. There is no term linear in because it was marginal under scaling, but it does receive a contribution when we integrate out the high momentum modes at order . This contribution is negative, which tells us the coupling is marginally irrelevant. (A repeat of the warning above: this is the opposite convention to quantum field theory where one flows in decreasing energy, rather than increasing length, which means that a marginally irrelevant interaction is usually said to have a positive beta function. )
The calculations above are very similar to the kind of loop integrals that you do in quantum field theory in dimensions. There are, however, some philosophical differences between the approaches.
In statistical mechanics, the field is, by construction, a coarse grained object: at the microscopic level, it dissolves into constituent parts, whether spins or atoms or something else. This has the practical advantage that we have no expectation that the statistical field theory will describe physics on arbitrarily short distance scales. In contrast, when we talk about quantum field theory in the context of high energy physics, it is tempting to think of the fields as “fundamental”, a basic building block of our Universe. We may then wish for the theory to make sense down to arbitrarily small distance scales.
This ambition leads to a subtly different viewpoint on renormalisation. In quantum field theory one must introduce a cut-off, as we have above, to render integrals finite. However, this cut-off is very often viewed as an artefact, one which we would ultimately like to get rid of and make sense of the theory as . The trouble is that the renormalised quantities – things that we’ve called and – typically depend on this cut-off. We saw this, for example, in (3.108). Often this makes it tricky to take the limit .
To avoid this problem, one makes the so-called bare couplings – things we’ve called and – depend on . This is not such a dumb thing to do; after all, these quantities were defined at the cut-off scale . The original game of renormalisation was to find a way to pick and such that all physical quantities remain finite as . It is by no means obvious that this is possible. Theories which can be rendered finite in this way are said to be renormalisable.
The high-energy approach to renormalisation predates the statistical physics approach and is now considered rather old-fashioned. The idea that a theory needs to make sense up to arbitrarily high energy scales smacks of hubris. The right way to view renormalisation – whether in statistical mechanics or in high energy physics – is through the renormalisation group procedure that has been our main focus in this chapter, in which one integrates out short wavelength modes to leave an effective long-distance theory.
Nonetheless, the high-energy approach to renormalisation has its advantages. Once one goes beyond the calculations described above, things are substantially easier with a high-energy viewpoint. You will learn more about these issues in the lectures on Advanced Quantum Field Theory.
We have learned that interaction is irrelevant for and relevant for , sweeping us away from the Gaussian fixed point. But we seem to be no closer to figuring out where we end up. All we know is that we’re not in Kansas anymore.
The difficulty is that we’re limited in what we can calculate. We can’t do the path integral exactly in the presence of interactions, and are forced to work perturbatively in the coupling . Yet, as we have seen, in dimension the RG flow increases , taking us to a regime where perturbation theory is no longer valid.
Nonetheless, the calculations that we did above do contain information about where we might expect to end up. But to see it, we have to do something rather dramatic. We will consider our theories not in dimensions, but in
dimensions where is a small number, much less than 1. Clearly, this is an odd thing to do. You could view it as an act of wild creativity or, one of utter desperation. Probably it is a little bit of both. But, as we shall see, it will give us the insight we need to understand critical phenomena.
First, we should ask whether it makes sense to work in a non-integer dimension. The lattice models that we started with surely need to be defined in dimension . Similarly, it was important for us that the free energy is local, meaning that it is written as an integral over space, and this too requires . However, by the time we get to the beta function equations, it makes mathematical, if not physical, sense to relax this and work in arbitrary . We can read off these beta functions from the RG equations (3.108): they are
where is the area of the sphere . We’ve introduced the dimensionless coupling . Note that the beta function for now includes a term linear in arising from the scaling.
It can be checked that the two-loop diagrams we neglected contribute only at order . This means that it’s consistent to truncate to the beta functions above. We could use the general formula for the area of a sphere, , but this will give corrections of order , so instead we simply use . Similar comments apply to . We’re left with
The novelty of these beta functions is that they have two fixed points. There is the Gaussian fixed point that we discussed before. And there is a new fixed point,
Since we’re working to leading order in , the solution is
This is the Wilson-Fisher fixed point. Importantly, when is small then the fixed point is also small, so our calculation is self-consistent (although, since we are in a fractional dimension, arguably unphysical!).
To understand the flows in the vicinity of the new fixed point, we write and . Linearising the beta functions, we find
where, as with all our other calculations, the entries in the matrix hold only up to .
The eigenvalues of a triangular matrix coincide with the diagonal entries. We see that this fixed point has one positive and one negative eigenvalue,
In other words, the Wilson-Fisher fixed point has one relevant direction and one irrelevant. The flows are shown in the figure.
We see that the epsilon expansion provides us with a global picture of the RG flows in dimensions. One can check that all other couplings are also irrelevant at the Wilson-Fisher fixed point. Crucially, the fixed point sits at small where our perturbative analysis is valid.
Now suppose that we increase . The Wilson-Fisher fixed point moves to higher , and our perturbative approach breaks down. Nonetheless, it is not unreasonable to suppose that the qualitative picture of the flows remains the same. Indeed, this is thought to happen. Because the Wilson-Fisher fixed point has just a single relevant operator, it means that we will generically end up there if we we’re willing to tune just a single parameter, namely .
It is now a simple matter to compute the critical exponents in the epsilon expansion. Recall from Section 3.2 that these are related to the scaling dimensions of various terms. The relevant direction away from the Wilson-Fisher fixed point is temperature, . Its dimension is determined by the way it scales as we approach the critical point, . But this is precisely the eigenvalue that we just computed.
The critical exponent , defined by , is then given by (3.76)
We can then use the hyperscaling relation , given in (3.77), to compute the critical exponent for the heat capacity
To compute the other critical exponents, we need to evaluate the anomalous dimension . As we mentioned briefly above, this is related to the diagram and turns out to be , which is higher order in the expansion. We then have, from (3.74),
Equations (3.78), (3.79) and (3.80) then give
where all expressions hold up to corrections of order .
Of course, our real interest lies in the system at , or . It would be in poor taste to simply plug in . But I know that you’re curious. Here’s what we get:
MF | 0 | 1 | 3 | 0 | ||
---|---|---|---|---|---|---|
0.17 | 4 | 0.58 | ||||
0.1101 | 0.3264 | 1.2371 | 4.7898 | 0.0363 | 0.6300 |
Our answers are embarrassingly close to the correct values given the dishonest method we used to get there. One can, however, make this approach more respectable. The expansion has been carried out to order . It is not a convergent series. Nonetheless, sophisticated resummation techniques can be used to make sense of it, and the resulting expressions give a fairly accurate account of the critical exponents.
The real power of the epsilon expansion, however, is more qualitative than quantitative; it usually – but not always – gives a reliable view of the structure of RG flows.
We have not yet discussed much about what happens in dimensions. Here the story is somewhat richer. The first hint of this can be seen in a simple analysis of the Gaussian fixed point, which shows that
This means that the Gaussian fixed point has an infinite number of relevant deformations since , for each , is relevant.
It turns out that, in contrast to , there are actually an infinite number of fixed points in . Roughly speaking, the fixed point can be reached from the Gaussian fixed point by turning on
Of course, as we’ve seen above, the RG flow is not quite so simple. When we turn on the coupling we will generate all other terms, including and and so on. To reach the fixed point, we should tune all such terms to zero as we flow towards the infra-red.
One can show that the fixed point has relevant operators: schematically, these can be thought of as although, as above, there will be mixing between these. By turning on the least relevant operator, one can flow from the fixed point to the fixed point.
The results above are not derived using the expansion which, unsurprisingly, is not much use in . Instead, they rely on something new which we will briefly describe in Section 3.6.
“After about a month of work [at General Atomic Corp] I was ordered to write up my results, as a result of which I swore to myself that I would choose a subject for research where it would take at least five years before I had anything worth writing about. Elementary particle theory seemed to offer the best prospects of meeting this criterion.”
Kenneth Wilson
Renormalisation first entered physics in the context of quantum field theory, with the need to make sense of the UV divergences that arise in quantum electrodynamics. The theory, developed by Schwinger, Feynman, Tomonaga, Dyson and others, amounts to finding a consistent way to subtract one infinity from another, leaving behind a finite answer. This method yields excellent agreement with experiment but is, in the words of Feynman, a “dippy process”, in which the infinities are not so much understood as swept under a very large rug.
The first hint of something deeper – and the first hint of a (semi)-group action – was seen in the work of Gell-Mann and Low in 1954. They realised that one could define an effective charge of the electron, where denotes the energy scale at which the experiment takes place. This interpolates between the physical charge, as , and the so called bare charge at high energies.
Meanwhile, throughout the 50s and 60s, a rather different community of physicists were struggling to understand second order phase transitions. It had long been known that Landau theory fails at critical points, but it was far from clear how to make progress, and the results that we’ve described in this course took several decades to uncover. In Kings College London, a group led by Cyril Domb stressed the importance of focussing on critical exponents; at Cornell University, Benjamin Widom showed that the relationships between critical exponents could be derived by invoking a scale invariance, albeit with little understanding of where such scale invariance came from; and at the University of Illinois, Leo Kadanoff introduced the idea of “blocking” in lattice models, a real-space renormalisation group in which one worked with successively coarser lattice models.
While many people contributed to these developments, the big picture, linking ideas from particle physics, statistical physics and condensed matter physics, is due mostly to…
Ken Wilson received his PhD in 1961, working with Murray Gell-Mann on an assortment of topics in particle physics that fed his interest in the renormalisation group. He went on to spend much of the 1960s confused, scrabbling to understand the physics of scale, first in quantum field theory and later in the context of critical phenomena. He wrote very few papers in this time, but his reputation was strong enough to land him postdocs at Harvard and CERN and later even tenure at Cornell. (Some career advice for students: the strategy of being a genius and not writing anything rarely leads to such success.)
The floodgates opened in 1971 when Wilson set out his grand vision of the renormalisation group and, with his colleague Michael Fisher, suggested the epsilon expansion as a perturbative method to compute critical exponents in a paper charmingly titled “Critical Exponents in 3.99 Dimensions”. Wilson used these methods to solve the “Kondo problem” in which an isolated spin, sitting in a bath of mobile electrons, exhibits asymptotic freedom, and he was among the first to understand the importance of numerical approaches to solve statistical and quantum field theories, pioneering the subject of lattice gauge theory. In 1982 he was awarded the Nobel prize for his contributions to critical phenomena.
There are many questions that we have not yet answered? How do we know the critical exponents in exactly? How do we know that there are an infinite number of fixed points in ? Why are the critical exponents in rational numbers while, in they have no known closed form? How are we able to compute the critical exponents to 5 significant figures?
The answers to all these questions can be found in the emergence of a rich mathematical structure at the critical point. As we’ve seen throughout these lectures, the basic story of RG ensures that physics at the critical point is invariant under scale transformations
(3.110) |
More surprising is the fact that the physics is invariant under a larger class of symmetries, known as conformal transformations. These transformation consist of any map
(3.111) |
for some function . Such conformal transformations have the property that they preserve angles between lines.
The equation (3.111) has obvious solutions, such as translations and rotations, for which . Furthermore, it is simple to see that the scaling (3.110) falls into the class of conformal transformations, with . However, it turns out that there is one further, less intuitive transformation that obeys this condition. This is known as the special conformal transformation and is given by
(3.112) |
parameterised by an arbitrary vector .
The first question that we should ask is: why are theories at the fixed points invariant under the larger group of conformal transformations, rather than just scale transformations? The answer to this, which goes somewhat beyond this course, involves a deeper understanding of the nature of the RG flows and hinges, crucially, on being able to construct a quantity which decreases monotonically along the flow. This quantity is, unhelpfully, called in dimensions, in dimensions and in dimensions, and the fact that it decreases monotonically is referred to as the -theorem, -theorem and -theorem respectively. Using this machinery, it is then possible to prove that scale invariance implies conformal invariance. (The proof is more clearcut in ; it relies on some extra assumptions in higher , and there is a general feeling that there is more to understand here.)
The existence of an extra symmetry (3.112) brings a newfound power to the study of fixed points. The translational symmetries, rotational symmetries, special conformal symmetries and single scale transformation combine to form the conformal group, which can be shown to be isomorphic to . All fields and, in particular, all correlation functions must fall into representations of this group, a fact which restricts their form. In recent years, our understanding of these representations has allowed new precision in the computation of critical exponents in dimensions. This programme goes by the name of the conformal bootstrap.
In dimensions, conformal symmetry turns out to be particularly powerful. The group of finite conformal transformations follows the pattern in higher dimensions, and is . However, something rather special happens if you look at infinitesimal transformations where one finds that many many more are allowed. In fact, there are an infinite number. This means that there is a powerful, infinite dimensional algebra, known as the Virasoro algebra, underlying conformal theories in dimensions. This places much stronger constraints on these fixed points, ultimately rendering many of them solvable without resorting to perturbation theory. This is the reason why the critical exponents are rational numbers which can be computed exactly. This is also what allows us to understand the structure of the infinite number of multi-critical fixed points described in Section 3.5.2.
Conformal field theory in dimensions is a vast subject which arises in many different areas of physics. Although originally developed to understand critical phenomena, it also plays an important role in the lectures on the Quantum Hall Effect and the lectures on String Theory, where you can find an introduction to the basics of the subject.