Wednesday, June 19, 2013
... //

Kenneth Wilson died from complications of lymphoma (a blood cancer) in Saco, Maine (where he and his wife were previously brought due to their love for kayaking) on Saturday, aged 77 years and 1 week. He received his Nobel Prize in 1982. His adviser was Murray Gell-Mann and students included Jackiw, Shenker, Peskin, and Ginsparg.

More importantly, he taught us about the concepts of effective field theories and the renormalization group that have explained why the renormalization works – and many other things. Many folks – a set that includes my PhD ex-adviser Tom Banks – classify Wilson's insights as the deepest advance of theoretical physics since the 1970s. Despite these experts' opinions, Wilson remained largely unknown to the public throughout his life.

The first talk of my life presented at a university different than my own that I ever gave was a 1998 talk at the Ohio State. Wilson has been there since 1988 and he could have arrived to the talk but (even though I was immensely interested in his presence) I have completely forgotten whether he actually has arrived. ;-)

I first met Wilson and talked to him during a lunch in the Society of Fellows overlapping with the 2005 Sidneyfest. He was smiling and satisfied and he was still thinking about physics although his most beloved newest theories seemed self-evidently silly – not only according to me but also according to some fellow Nobel prize winners – and he wasn't quite following the ongoing cutting-edge theoretical research. I also knew a younger Slovak lady (Martina M. Brisudova) who was his recent collaborator.

Kenneth Wilson is the father of the Wilson loop, the path-ordered exponential over a closed loop that counts the trace of the monodromy,\[

But more importantly, he became the main guy behind the Renormalization Group (RG). Physicists had learned the playful and clever tricks of renormalization but they didn't quite understand where it came from and some of them had doubts whether it should have been trusted at all.

Remotely related: On Thursday, 5 pm Prague Summer Time i.e. 11 am Boston Daylight Time, there will be a Google Hangout with top HEP phenomenologist John Ellis about SUSY on youtube.com/CERNTV. Use @CERN #askcern on Twitter to ask questions. Incidentally, the first tetraquark (a pair of mesons stuck together) called \(Z_c(3900)\) was discovered at Belle as well as Beijing EPC, see e.g. Nature.

The Renormalization Group with its related machinery and terminology including effective field theories, relevant and irrelevant interactions, fixed points, and so on has eliminated all the doubts, unmasked the power that makes the renormalization procedures consistent and successful with a remarkable clarity, and gave us a modern understanding of what quantum field theory actually means (some people say that we are still waiting for analogous insights about the "true nature" of string theory). Wilson achieved these things in 1971-1974, building on the shoulders of Freeman Dyson's systematic theory of the old renormalization methodology from 1949 and Leo Kadanoff's 1966 ideas about the "block spin renormalization group".

What does this Wilsonian theory (some people could call it "Wilsonian philosophy" but this label doesn't reduce its robustness and importance in physics at all) say?

It says that quantum field theories (and similarly models in statistical physics that are mathematically analogous) should not be viewed as the final theories of everything but just as approximate theories that describe all objects and phenomena whose characteristic length scales are (much) longer than some \(L\) or, equivalently, whose energies are (much) lower than \(E\).

An important fact is that such a "restriction of the original theory", possibly a final theory, is possible at all. Why is it possible? Because we can explicitly construct it. Assuming that your "more complete" theory admits a formulation in terms of Feynman's path integral, we may define\[

\] On the right hand side, we are using a theory with the action \(S_\Lambda\) and this theory is supposed to work for all energies/momenta up to \(\Lambda\) which is very high. You may imagine this parameter to be infinite if you haven't thought about theories with a restricted domain of validity before.

All the calculable probability amplitudes are given by the Feynman path integral which is an infinite-dimensional integral over all field modes with various momenta. The key observation is that this integral may be reorganized in such a way that we first integrate it over the higher-energy modes, e.g. – in the formula above – modes with \(\Lambda' \leq p \leq \Lambda\). In this way, we obtain a function that only depends on the low-energy field modes, \(p\leq \Lambda'\), and the integral over these field modes can be done at the end.

A funny thing is that the function we integrate at the end only depends on the low-energy field modes – because the higher-energy field modes have been "integrated out" which means that they have been "integrated over" which excluded them "out of the list of variables upon which our favorite/remaining action on the left hand side \(S_{\Lambda'}\) depends"). Still, this simplified function is totally sufficient to calculate arbitrary correlators etc. of the low-energy field modes (and scattering amplitudes for particles at low energies, among related things) as long as we "integrated out" the high-energy quanta properly and accurately.

The function that only depends on the low-energy quanta defines what we call the "effective field theory". Because its action doesn't depend on the high-energy quanta at all, this "effective field theory" will also generally become independent of any particles, fields, interactions, and laws of physics that only influence the very-short-distance or very-high-energy physical phenomena. We don't need to know the quarks to study atomic physics (or chemistry) and the Wilsonian "integrating things out" quantitatively realizes the same general idea in the technical framework of quantum field theories.

(You should bring your mind to the right mood by checking one of the interactive Flash animations showing the Universe at various length scales. Wilson effectively tells us to study the scales independently.)

So different theories valid for all distance scales, including the very short ones, may produce the same - or nearly the same – effective field theories for the low-energy modes. They may just imply the same spectrum of particles or fields at low energies and because their interactions are rather constrained (the space of effective field theories obeying certain extra conditions is rather small or exclusive), the interactions may agree, too.

Celebrations of the 1982 Nobel prize at Cornell. He looks very young among his colleagues – we're used to young people celebrating old men's Nobel prize – but he was already 46 on the picture above.

This was the first, more general part of the Wilsonian ideas: it's a good idea to separate physics to the physics at various scales. Short-distance physics affects long-distance physics that is derived from it; but the relationship doesn't hold in the opposite direction because short-distance physics is often left undetermined if we only know its long-distance manifestations.

The second part of Wilson's important contributions is a whole industry of methods that tell us how the effective field theories differ from the original ones in the case that the original ones are also quantum field theories, and we could even say that they are effective field theories as well but ones with a higher \(\Lambda\) and how the space of possible effective field theories may be parameterized.

When I wrote the only big displayed equation above, I encouraged you to imagine that \(\Lambda\), the highest scale at which the original theory was valid, was infinite while \(\Lambda'\), the highest scale where the effective (derived) theory is applicable, is much smaller. However, the real technical power of the Renormalization Group shows up when the scales \(\Lambda\) and \(\Lambda'\) are actually very close to each other:\[

\Lambda' = \Lambda (1-\varepsilon)

\] Here, \(\varepsilon\) is an infinitesimal positive number. In this case, the partial integration in the Feynman path integral is the integration over a thin shell of field modes \(\phi(p)\) whose momenta (their magnitude) belong to a very narrow interval\[

\Lambda(1-\varepsilon) \leq p \leq \Lambda.

\] In other words, we are just trying to lower the scale \(\Lambda\) by an infinitesimal amount. This changes the original quantum field theory to something else but because the change we have made is apparently "infinitesimal", the change of the quantum field theory should be infinitely small, too.

In fact, the derived effective field theory will be a theory of the very same kind as the original one but the values of the parameters – masses of particles and coupling constants – will be changed by an infinitesimal amount. We may always interpret the lowering of the value of \(\Lambda\) as a "transformation" and these transformations may be composed associatively. There is also an identity transformation (keep \(\Lambda\) and therefore the quantum field theory intact) so we may say that these transformations that lower the values of \(\Lambda\) form a group.

Well, more precisely, we have said that the transition from a more complete theory with a higher \(\Lambda\) to an effective field theory with a lower \(\Lambda'\) is irreversible because this procedure is "forgetting" some particles and interactions that only mattered at high energies. Because of this irreversibility, the transformations lowering the values of \(\Lambda\) don't admit any inverse transformations. An almost group without the condition that the inverse elements exist is called a semigroup but because physicists would think that the term Renormalization Semigroup is awkward, hard to pronounce, and dominated by mathematicians' nitpickiness, they use the term Renormalization Group. The (not quite) group elements are still the (associative) transformations reducing the value of the \(\Lambda\), the maximum energy scale at which the effective theory works.

The procedure of lowering \(\Lambda'\) has some impact on the parameters of the effective field theory. This effect may be calculated by Feynman diagrams (at least perturbatively) in which the internal lines are only integrated over a small interval or shell of allowed momenta and energies. When you do such a thing, you will find out that the couplings "run". They depend on \(\Lambda\). (When you discuss the same kind of changes of all the parameters and perhaps even more qualitative changes of the whole theory that make the theory "run", the right verb is that we are "flowing the theory to the infrared".) The most important and perhaps the most typical functional dependence that appears in this running is the logarithmic running, something like (approximately, up to 1-loop diagrams)\[

\] where the constant prefactor \(B\) is related to the so-called \(\beta\)-function, the "rate" by which the coupling constant changes with \(\Lambda\). Similar and perhaps more complicated "RG equations" are used to study how the parameters evolve from the high-energy scale to a low-energy scale. In particular, these "running coupling" calculations are totally essential to discuss the gauge coupling unification (convergence of the "fine-structure constants" of the three factors of the Standard Model gauge group to a common value at a high energy scale) in grand unified theories and for many similar applications. It's important to realize that as long as we identify the couplings with finite numbers that really correspond to some processes at a given energy, they are allowed to run.

If you want to use the RG methods to understand why the old renormalization methods – already used since the 1940s – work, it is a good idea to "map" the space of possible effective theories with a given spectrum and with some fixed value of \(\Lambda\). If these theories form an \(n\)-dimensional space, it must be possible to deform each of them to get to a nearby effective field theory. These deformations may in turn be realized by adding a term (operator) to their Lagrangian.

For an effective field theory, you want to classify all possible deformations. They may be divided to relevant ones, marginal ones (the "unlikely", generically measure-zero border case), and irrelevant ones according to their influence on the very low-energy physics. In general, the relevant deformations are those whose effect is increasingly important as you move from high energies to low energies; the rule is reverted for the irrelevant ones and the effect remains equally strong at all scales for the marginal ones.

The most reductionist treatment of the perturbatively known quantum field theories such as QED or the Standard Model presents all of them as deformations of a "Gaussian fixed point". The adjective "Gaussian" means that the integrand of the path integral is Gaussian i.e. that the action is free (at most bilinear); there also exist non-Gaussian (interacting) fixed points but they're harder to be found. The deformations are all the interactions we are adding. The term "fixed point" refers to the theory's being unchanged under the renormalization group flows i.e. its being independent of \(\Lambda\): fixed points are nothing else than scale-invariant theories – the most important lighthouses in the landscape of effective field theories according to the RG methods to map this landscape.

The deformations may be roughly identified with the extra terms in the Lagrangian that you might add. You will find out that the relevant ones are those whose coefficients have units of \({\rm mass}^n\), positive powers of mass, while the irrelevant ones have negative powers of mass. You will only find a finite number of relevant deformations but an infinite number of irrelevant ones – the latter are the "non-renormalizable interactions" (also essentially equivalent to what physicists call "higher-dimension operators"), such as \[

\delta S = L^4\cdot (F_{\mu\nu}F^{\mu\nu})^2

\] in quantum electrodynamics where \(L\) is some parameter with the units of length.

Before Wilson, non-renormalizable interactions could have been interpreted as the ultimate blasphemies, extra terms that immediately throw us to a hell of inconsistencies (an infinite hell because there are infinitely many such terms we may add), something that we shouldn't even think about. Wilson's appraisal of their status is different. They're OK, you may actually add them, but they're "irrelevant" because their effect on the effective field theory below the scale \(\Lambda'\) becomes negligible if this scale is much smaller than the original one, \(\Lambda'\ll \Lambda\).

If you generate an irrelevant interaction in an effective field theory from the "integrating out" of some field modes, the typical magnitude of the parameter \(L\) above will be of order \(1/\Lambda\), i.e. linked to the very high-energy scale where the source of the interaction resides. This is why the effect of such a higher-dimension operator will be negligible around the low energy scale \(\Lambda'\) because the coefficient\[

L^4 \sim \frac{1}{\Lambda^4} \ll \frac{1}{\Lambda^{\prime 4}}

\] is much smaller, by the factor of \((\Lambda'/\Lambda)^n\) with some positive exponent \(n\), in this case \(n=4\), than the typical size of the coefficient that you would have to expect (by dimensional analysis) if this interaction were as important as some relevant or marginal ones at the energy scale close to \(\Lambda'\).

Once again, instead of being "immediate superstrong devils and killers of consistency", irrelevant interactions were reclassified as effectively harmless bugs. The higher the gap is between the low energy that you experimentally probe and the high energy scale where the irrelevant term originates, the more negligible they will be. Despite the small coefficient, they may still sometimes be important, especially if they generate rare processes that can't be caused by any relevant, marginal or otherwise "normally strong" interactions.

The marginal interactions are in between. For example, the fine-structure constant \(\alpha\sim 1/137.036\) is dimensionless which means that the characteristic strength of the electromagnetic interactions is linked to a marginal deformation. Well, because this fine-structure logarithmically runs, it's actually not exactly marginal. These couplings have "anomalous dimensions" – the exponents have corrections proportional to \(\alpha\) themselves. So the fine-structure constant only looks dimensionless classically; quantum mechanically, the corresponding coefficients have the units of a fractional power of the energy that is just close but not equal to the power derived classically.

(If you want exactly marginal deformations, you demand the quantum correction to the classical dimension – the anomalous dimension – to vanish exactly as well. This rarely occurs by chance and almost all important examples we know, at least for \(d\gt 2\), are supersymmetric theories. Supersymmetry likes to guarantee similar cancellations. We also know important interacting supersymmetric theories that are nevertheless fixed points, i.e. exactly scale invariant. The \(\NNN=4\), \(d=4\) gauge theory is the most celebrated example while a non-Lagrangian six-dimensional \((2,0)\) theory is its much less well-known cousin.)

Ascania: Supersymmetry.

Such RG methods may also convince you that it doesn't matter which kind of a regularization – brute cutoffs, Pauli-Villars, dimensional regularization etc. – you use. The Wilsonian idea is that you focus on the space of effective theories i.e. those that are directly useful for the predictions of doable low-energy experiments. This space of theories – defined to be "almost directly relevant for the observations" – may be shown to exist and to have a certain dimensionality or allowed deformations and there may be many ways how this space is described or parameterized. These methods must ultimately differ by a redefinition of variables only. Whatever you can do with one regularization technique or renormalization scheme, must be translatable to another.

The "integrating out" is the key technique that allows us to translate the properties of the high-energy quantum field theory – something that may be rather directly linked to a more fundamental theory that doesn't have to be a local quantum field theory, especially to string theory – into the properties of the low-energy effective field theories that is almost immediately usable to describe the doable observations.

It's important that this translation – and the running of the couplings or flowing of the theories etc. – exists at all and it is not an identity transformation. It's important that the low-energy effective field theory is independent of many or most details of the high-energy physics. The previous sentence is pretty much equivalent to an observation from a different angle, namely that the behavior of quantum field theories (and even other high-energy starting points such as string theory) at low energies tends to be "universal". These possible low-energy behaviors may be discussed separately from the dynamics at high energies or short distances.

So what about the infinities that the old renormalization uses (and has to cancel) all the time? In the renormalization group philosophy, you may imagine that these are finite numbers that depend on a high energy scale \(\Lambda\). These terms have to cancel by definition if our task is to study effective field theories i.e. descriptions that are independent of the physics above the high energy scale \(\Lambda\). In particular, the effective field theory has to be independent of \(\Lambda\) itself.

The cancellation of the divergences is no magic or blasphemy anymore. Wilson has shown that this cancellation pretty much tautologically follows from the very task we outlined for ourselves – the task is to study the observable low-energy phenomena which effectively means to study the effective field theory for a physical system (or the possible effective field theories for a class of systems). Because of this independence, one may also get rid of some contrived artifacts linked to a particular finite value of \(\Lambda\) and study the limit \(\Lambda\to\infty\) in which the cancelled terms are "strictly" infinite. It's just a natural limit that makes the unimportance of the physics at the high energy scale more self-evident.

The Wilsonian approach leads to a revision of many ideas about naturalness, the real problems with non-renormalizable theories, and more. Whether a theory is natural or not should be decided according to the values of the parameters at the high, fundamental energy scale; the values at low energies are their consequence. However, it may often be hard for a high-energy theory to "flow" to a realistic or semirealistic theory at low energies, e.g. to preserve any light particles that survive at all (if there are no particles lighter than \(\Lambda'\), the "integrating out" may leave us with no degrees of freedom at all; the path integral becomes a boring constant because there are no variables left). The infinities themselves aren't a problem because you may always imagine that those numbers are finite; the real problem of the non-renormalizable interactions is that there are infinitely many of them whose coefficients have to be adjusted which makes the theory unpredictive for the phenomena near \(\Lambda\).

All these insights were found independently of string theory and, effectively ;-), before string theory. And Ken Wilson wasn't even a string theorist at any point of his life (sorry, I don't count his strings on a lattice). Still, pretty much all the people who talk about nonsensical things such as "competing theories", "loop quantum gravity", and so on misunderstand most of the insights about the renormalization group – even the general comments above. Their beliefs about the character and right interpretation of renormalization techniques are stuck somewhere in the 1940s (especially because of the patently obsolete opinion that the real challenge when it comes to UV divergences is to get rid of divergent integrals). In this sense, these "anti-string-theorists" misunderstand not only the physics of the last 40 years but also the physics of the last 70 years. They're just hopeless.

The name of Ken Wilson in this very form has appeared in more than 20 older TRF blog entries. RIP.

snail feedback (21)
:

reader
RF
said...

Dear Dr Luboš Motl,

It is inappropriate off topic comment regardless of your mourning post. But I expected your attention to my post at http://physics.stackexchange.com/questions/68508/mass-term-in-the-minkowski-metric-and-an-oscillation-model-in-lorentz-groupHowever it is already closed before the chance of receiving your suggestion as I tagged string-theory on that post. If you have any idea to remark about the paper in the post, please find that there is another post athttp://www.physicsforums.com/showthread.php?t=697709

At the end, I always enjoy your blogs for so long time even though I could catch up less than quarter of it. As many TRF readers say, I alsofeel like I am a student of string theory when reading ;)

Dear Ryoji, thanks for your interest but apologies, I would also close your "question" if I were a moderator. It doesn't seem to be a question at all. Moreover, even as a non-question, it doesn't make any sense. Do you want to revise 1905 special relativity? Why? Isn't it better to learn it? It is not that hard. If you think you have some arguments for a revision of relativity, I was surely unable to find it.

Moreover, I believe that you must see that you are doing something really inappropriate. Without any glimpse of red face, you inform us that you tagged question with the string-theory tag to increase your "chance of receiving my suggestion". That's humbling. But holy cow, can't you see that your question - even if we called it a question which it's not - has absolutely nothing to do with string theory? Do you understand that the tags should actually describe the topic of the question and when they don't, it's bringing noise and bullshit to the server? Don't you understand that you were really abusing the tags and it's unproductive if not immoral to do such things?

Even if you ignore all these things, it's unproductive even as a way to get my attention. I don't follow any tags in a selective way. Sometimes I look at some questions, at least their titles, sometimes perhaps I don't, but the tags really don't affect this selection in my case.

Dirac has written many times that out interaction term jA is wrong because of wrong concepts behind it; that's why the theory gives wrong results and we need to "repair" them. We should search for a better theory formulation, conceptually and mathematically.

K. Wilson says the interaction term jA is, of course, right, but we should not use it because it is wrong. Instead we should cut it off, renormalize constants, etc., etc. So everything is already OK. What an impressive conceptual progress!!!

Ohio State University hired Dr. Wilson because they wanted his wife to run the new Ohio Supercomputer Center. It was a two-fer.

Ohio had lost out in the NSF competition to be a member of the Midwest regional center, so the State built its own on the OSU campus. The State has continued its investment in the Center, and it is nearly, but not quite, state of the art. It is available essentially free of charge to any faculty member in Ohio with a reasonable proposal.

Dr. Wilson was OSU's only Noble laureate, and his was a grace note on our faculty.

Lubos, thanks for this excellent and very readable post that describes Ken Wilson's contributions so clearly. In reading it, I could not help thinking that the post itself was an example of RNG behavior, that is, you were "flowing the theory to the infrared" by integrating out higher "energy" technicalities to provide a low "energy" description that was more broadly accessible. I suppose this too is likely to be a semigroup ;-)

You said at P. SE, that there is no physics department at your university. So if you are seriously interested in fundamental physics and even want to contribute to it, you should change the university and properly study physics first.

Wilson was the kind of scientist that I most admire. He was not focused on self-promotion but on science itself. I am sure he had a very satisfying life amid all the politics and self-aggrandizement surrounding him.

I sometimes leave it out when asking even when I think I could use it, for example when my question is potentially too confused or dumb... In addition, I have used enough of them and overdoing it makes me look too much like the geek I actually am :-P :-D ;-) ...

yThanks for this very very nice overview of the key points of renormalization.

The point of view Wilson has given us I find so much more beautiful and intuitively understandable, than the old renormalization which looks some kind of awkward to me. And I keep being excited upon learning more and more how these ideas and point of view bring together different things in theoretical physics. From reading your comments about the irreversibility of the coarse graining step I think renormalization and nonequilibrium statistical physics must be closely related and I have found and printed out a relatively new paper discussing this indeed :-)

It is sad, that Wilson is now no longer among us, but his great insights will remain.

Dear Dilaton, the old renormalization procedures are still being done after Wilson in pretty much the same way as before! (At least when one does perturbative calculations.) He "only" changed the way how we think about these methods.

but Lumo often does the converse too. When you think you tead an article just about some decent statistical mechanics or QM, he always smuggles in some very high energy degrees of freedom, at least somewhere in the text :-D

Dear Jan, thanks for your interest. Scale invariance is as meaningful in a d-dimensional QFT as it is in statistical mechanics at some temperature describing a system with d spatial dimensions and 1 time.

For example, Maxwell's electrodynamics in the vacuum is scale-invariant because it has no dimensionful parameters. Take some electromagnetic waves/fields, magnify them 7 times, and they will still solve Maxwell's equations.

However, there also exist fixed point QFTs that are interacting theories. QCD is classically scale-invariant, too. Quantum corrections often violate the precise scale invariance and QCD is an example of that. However, there are other QFTs that are interacting and exactly scale-invariant. The quantum fluctuating fields in these theories are fractal-like and really mathematically resemble the critical opalescence and other things you know from statistical physics.

Note that in QFTs, we *need* zero temperature for scale invariance. A nonzero temperature 1/beta = T is the inverse periodicity of the Euclidean time, a dimensionful parameter that picks a preferred scale and violates the scale invariance even if it was there to start with.