# Romeo and Juliet, through a Wormhole

I spent last week at the Perimeter Institute in Waterloo, Ontario. Undoubtedly one of the highlights was Juan Maldenena’s keynote on resolving black hole paradoxes using wormholes. Matt’s review of the talk below is well worth a read!

Perimeter is hosting this year’s Mathematica Summer School on Theoretical Physics. The school is a mix of lectures on a topic in physics (this year, the phenomenon of quantum entanglement) and tips and tricks for using the symbolic calculation program Mathematica.

Juan Maldacena is one of the lecturers, which gave me a chance to hear his Romeo and Juliet-based explanation of the properties of wormholes. While I’ve criticized some of Maldacena’s science popularization work in the past, this one is pretty solid, so I thought I’d share it with you guys.

You probably think of wormholes as “shortcuts” to travel between two widely separated places. As it turns out, this isn’t really accurate: while “normal” wormholes do connect distant locations, they don’t do it in a way that allows astronauts to travel between them, Interstellar-style. This can be illustrated with something called a Penrose…

View original post 626 more words

# Correlation Functions in Cosmology – What Do They Measure?

The cosmic microwave background (CMB) is a key observable in cosmology.  Experimentalists can precisely measure the temperature of microwave radiation left over from the big bang. The data shows very small differences in temperature across the sky. It’s up to theorists to figure out why!

The most popular explanation invokes a scalar field early in the universe. Quantum fluctuations in the field are responsible for the classical temperature distribution we see today. This argument, although naively plausible, requires some serious thought for full rigour.

Talks by cosmologists often parrot the received wisdom that the two-point correlation function of the scalar field can be observed on the sky. But how exactly is this done? In this post I’ll carefully explain the winding path from theory to observation.

First off, what really is a correlation function? Given two random variables $X$ and $Y$ we can (roughly speaking) determine their correlation as

$\mathbb{E}(XY)$

Intuitively this definition makes sense – in configurations where $X$ and $Y$ share the same sign there is a positive contribution to the correlation.

There’s another way of looking at correlation. You can think of it as a measure of the probability that for any random sample of $X$ there will be a value of $Y$ within some given distance. Hopefully this too feels intuitive. It can be proved more rigorously using Bayes’ theorem.

This second way of viewing correlation is particularly useful in cosmology. Here the random variables are usually position dependent fields. The correlation then becomes

$\langle \chi(x)\chi(y) \rangle$

where the average is over the whole sky with the direction of the vector $x- y$ fixed. The magnitude of this vector provides a natural distance scale for the probabilistic interpretation of correlation. We see that the correlation is an avatar for the lumpiness of the distribution at a particular distance scale!

Now let’s focus on the CMB. The temperature fluctuations are defined as the percentage deviation from the average temperature at each point on the sky. Mathematically we write

$\delta T / T (\hat{n})$

where $\hat{n}$ defines a point on the unit $2$-sphere. We want to relate this to theoretical predictions. Given our discussion above, it’s not surprising that our first step is to compute the correlation function

$C(\theta) = \displaystyle \langle \frac{\delta T}{ T}(\hat{n}_1) \frac{\delta T}{T}(\hat{n}_2)\rangle$

where the average is over the whole sky with the angle $\theta$ between $\hat{n}_1$ and $\hat{n}_2$ fixed. This average doesn’t lose any physical information since there’s no preferred direction in the sky! We can conveniently encode the correlation function using spherical harmonics

$\delta T / T = \sum a_{l,m} Y_{l,m}$

The coefficients $a_{l,m}$ are known as the multipole moments of the temperature distribution. Substituting this in the correlation function definition we obtain

$C(\theta) = \sum C_l P_l (\cos \theta)$

where $C_l = \sum_m |a_{l,m}|^2$. We’re almost finished with our derivation! The final step is to convert from the correlation function to it’s momentum space representation, known as the power spectrum. With a little work, you can show that the power at multipole number $l$ is given by

$l(l+1)C_l$

This is exactly the quantity we see plotted from sky map data on graphs comparing inflation theory to experiment!

From the theory perspective, this quantity is fairly easy to extract. We must compute the power spectrum of the primordial fluctuations of the inflation field. This is merely a matter of quantum field theory, albeit in de Sitter spacetime. Perhaps the most comprehensive account of this procedure is provided in Daniel Baumann’s notes.

Without going into any details, it’s worth mentioning a few theoretical models. The simplest option is to have a massless free inflaton field. This gives a scale-invariant power spectrum, which is almost correct! Adding mass corrects this result, providing fluctuations in the power spectrum. This is a better approximation, but has been ruled out by Planck data.

Clearly we need a more general potential. Here’s where the fun starts for cosmologists. The buzzwords are effective field theory, string inflation, non-Gaussianity and multiple fields! But that’ll have to wait for another blog post.

Written at the Mathematica Summer School 2015, inspired by Juan Maldecena’s lecture.

# What Gets Conserved at Vertices in Feynman Diagrams?

The simple answer is – everything! If there’s a symmetry in your theory then the associated Noether charge must be conserved at a Feynman vertex. A simple and elegant rule, and one of the great strengths of Feynman’s method.

Even better, it’s not hard to see why all charges are conserved at vertices. Remember, every vertex corresponds to an interaction term in the Lagrangian. These are automatically constructed to be Lorentz invariant so angular momentum and spin had better be conserved. Translation invariance is built in by virtue of the Lagrangian spacetime integral so momentum is conserved too.

Internal symmetries work in much the same way. Color or electric charge must be conserved at each vertex because the symmetry transformation exactly guarantees that contributions from interaction terms cancel transformations of the kinetic terms. If you ain’t convinced go and check this in any Feynman diagram!

But watch out, there’s a subtlety! Suppose we’re interested in scalar QED for instance. One diagram for pair creation and annihilation looks like

Naively you might be concerned that angular momentum and momentum can’t possibly be conserved. After all, don’t photons have spin and mass squared equal to zero? The resolution of this apparent paradox is provided by the realization that the virtual photon is off-shell. This is a theorist’s way of saying that it doesn’t obey equations of motion. Therefore the usual restrictions from symmetries do not apply to the virtual photon! Thinking another way, the photon is a manifestation of a quantum fluctuation.

Erratum: a previous version of this article erroneously claimed that Noether’s second theorem is related to Ward identities that guarantee gauge invariance at the quantum level. This is not the case, to our knowledge. Indeed, the Ward identity is a statement about averaging over field configurations, which necessarily depends on the behaviour of the path integral measure, a quantity that Noether never concerned herself with! Interestingly, there is a connection between the second theorem and large residual gauge symmetries, as pointed out in https://arxiv.org/abs/1510.07038.

# Conference Amplitudes 2015 – Don’t Stop Me Now!

All too soon we’ve reached the end of a wonderful conference. Friday morning dawned with a glimpse of perhaps the most impressive calculation of the past twelve months – Higgs production at three loops in QCD. This high precision result is vital for checking our theory against the data mountain produced by the LHC.

It was well that Professor Falko Dulat‘s presentation came at the end of the week. Indeed the astonishing computational achievement he outlined was only possible courtesy of the many mathematical techniques recently developed by the community. Falko illustrated this point rather beautifully with a word cloud.

As amplitudeologists we are blessed with a incredibly broad field. In a matter of minutes conversation can encompass hard experiment and abstract mathematics. The talks this morning were a case in point. Samuel Abreu followed up the QCD computation with research linking abstract algebra, graph theory and physics! More specifically, he introduced a Feynman diagram version of the coproduct structure often employed to describe multiple polylogs.

Dr. HuaXing Zhu got the ball rolling on the final mini-session with a topic close to my heart. As you may know I’m currently interested in soft theorems in gauge theory and gravity. HuaXing and Lance Dixon have made an important contribution in this area by computing the complete $2$-loop leading soft factor in QCD. Maybe unsurprisingly the breakthrough comes off the back of the master integral and differential equation method which has dominated proceedings this week.

Last but by no means least we had an update from the supergravity mafia. In recent years Dr. Tristan Dennen and collaborators have discovered unexpected cancellations in supergravity theories which can’t be explained by symmetry alone. This raises the intriguing question of whether supergravity can play a role in a UV complete quantum theory of gravity.

The methods involved rely heavily on the color-kinematics story. Intriguingly Tristan suggested that the double copy connection because gauge theory and gravity could form an explanation for the miraculous results (in which roughly a billion terms combine to give zero)! The renormalizability of Yang-Mills theory could well go some way to taming gravity’s naive high energy tantrums.

There’s still some way to go before bottles of wine change hands. But it was fitting to end proceedings with an incomplete story. For all that we’ve thought hard this week, it is now that the graft really starts. I’m already looking forward to meeting in Stockholm next year. My personal challenge is to ensure that I’m among the speakers!

Particular thanks to all the organisers, and the many masters students, PhDs, postdocs and faculty members at ETH Zurich who made our stay such an enjoyable and productive one!

# Conference Amplitudes 2015 – Air on the Superstring

One of the first pieces of Bach ever recorded was August Wilhelmj’s arrangement of the Orchestral Suite in D major. Today the transcription for violin and piano goes by the moniker Air on the G String. It’s an inspirational and popular work in all it’s many incarnations, not least this one featuring my favourite cellist Yo-Yo Ma.

This morning we heard the physics version of Bach’s masterpiece. Superstrings are nothing new, of course. But recently they’ve received a reboot courtesy of Dr. David Skinner among others. The ambitwistor string is an infinite tension version which only admit right-moving vibrations! At first the formalism looks a little daunting, until you realise that many calculations follow the well-trodden path of the superstring.

Now superstring amplitudes are quite difficult to compute. So hard, in fact, that Dr. Oliver Schloterrer devoted an entire talk to understanding particular functions that emerge when scattering just  $4$ strings at next-to-leading order. Mercifully, the ambitwistor string is far more well-behaved. The resulting amplitudes are rather beautiful and simple. To some extent, you trade off the geometrical aesthetics of the superstring for the algebraic compactness emerging from the ambitwistor approach.

This isn’t the first time that twistors and strings have been combined to produce quantum field theory. The first attempt dates back to 2003 and work of Edward Witten (of course). Although hugely influential, Witten’s theory was esoteric to say the least! In particular nobody knows how to encode quantum corrections in Witten’s language.

Ambitwistor strings have no such issues! Adding a quantum correction is easy – just put your theory on a donut. But this conceptually simple step threatened a roadblock for the research. Trouble was, nobody actually knew how to evaluate the resulting formulae.

Nobody, that was, until last week! Talented folk at Oxford and Cambridge managed to reduce the donutty problem to the original spherical case. This is an impressive feat – nobody much suspected that quantum corrections would be as easy as a classical computation!

There’s a great deal of hope that this idea can be rigorously extended to higher loops and perhaps even break the deadlock on maximal supergravity calculations at $7$-loop level. The resulting concept of off-shell scattering equations piqued my interest – I’ve set myself a challenge to use them in the next 12 months!

Scattering equations, you say? What are these beasts? For that we need to take a closer look at the form of the ambitwistor string amplitude. It turns out to be a sum over the solutions of the following equations

$\sum_{i\neq j}\frac{s_{ij}}{z_i - z_j}=0$

The $s_{ij}$ are just two particle invariants – encoding things you can measure about the speed and angle of particle scattering. And the $z_i$ are just some bonus variables. You’d never dream of introducing them unless somebody told you to! But yet they’re exactly what’s required for a truly elegant description.

And these scattering equations don’t just crop up in one special theory. Like spies in a Cold War era film, they seem to be everywhere! Dr. Freddy Cachazo alerted us to this surprising fact in a wonderfully engaging talk. We all had a chance to play detective and identify bits of physics from telltale clues! By the end we’d built up an impressive spider’s web of connections, held together by the scattering equations.

Freddy’s talk put me in mind of an interesting leadership concept espoused by the conductor Itay Talgam. Away from his musical responsibilities he’s carved out a niche as a business consultant, teaching politicians, researchers, generals and managers how to elicit maximal productivity and creativity from their colleagues and subordinates. Critical to his philosophy is the concept of keynote listening – sharing ideas in a way that maximises the response of your audience. This elusive quality pervaded Freddy’s presentation.

Following this masterclass was no mean feat, but one amply performed by my colleague Brenda Penante. We were transported to the world of on-shell diagrams – a modern alternative to Feynman’s ubiquitous approach. These diagrams are known to produce the integrand in planar $\mathcal{N}=4$ super-Yang-Mills theory to all orders! What’s more, the answer comes out in an attractive $d \log$ form, ripe for integration to multiple polylogarithms.

Cunningly, I snuck the word planar into the paragraph above. This approximation means that the diagrams can be drawn on a sheet of paper rather than requiring $3$ dimensions. For technical reasons this is equivalent to working in the theory with an infinite number of color charges, not just the usual $3$ we find for the strong force.

Obviously, it would be helpful to move beyond this limit. Brenda explained a decisive step in this direction, providing a mechanism for computing all leading singularities of non-planar amplitudes. By examining specific examples the collaboration uncovered new structure invisible in the planar case.

Technically, they observed that the boundary operation on a reduced graph identified non-trivial singularities which can’t be understood as the vanishing of minors. At present, there’s no proven geometrical picture of these new relations. Amazingly they might emerge from a 1,700-year-old theorem of Pappus!

Bootstraps were back on the agenda to close the session. Dr. Agnese Bissi is a world-expert on conformal field theories. These models have no sense of distance and only know about angles. Not particularly useful, you might think! But they crop up surprisingly often as approximations to realistic physics, both in particle smashing and modelling materials.

Agnese took a refreshingly rigorous approach, walking us through her proof of the reciprocity principle. Until recently this vital tool was little more than an ad hoc assumption, albeit backed up by considerable evidence. Now Agnese has placed it on firmer ground. From here she was able to “soup up” the method. The supercharged variant can compute OPE coefficients as well as dimensions.

Alas, it’s already time for the conference dinner and I haven’t mentioned Dr. Christian Bogner‘s excellent work on the sunrise integral. This charmingly named function is the simplest case where hyperlogarithms are not enough to write down the answer. But don’t just take it from me! You can now hear him deliver his talk by visiting the conference website.

Conversations

I’m very pleased to have chatted with Professor Rutger Boels (on the Lagrangian origin of Yang-Mills soft theorems and concerning the universality of subheading collinear behaviour) and Tim Olson (about determining the relative sign between on-shell diagrams to ensure cancellation of spurious poles).

Note: this post was originally written on Thursday 9th July but remained unpublished. I blame the magnificent food, wine and bonhomie at the conference dinner!

# Conference Amplitudes 2015 – Integrability, Colorful Duality and Hiking

The middle day of a conference. So often this is the graveyard slot – when initial hysteria has waned and the final furlong seems far off. The organisers should take great credit that today was, if anything, the most engaging thus far! Even the weather was well-scheduled, breaking overnight to provide us with more conducive working conditions.

Integrability was our wake-up call this morning. I mentioned this hot topic a while back. Effectively it’s an umbrella term for techniques that give you exact answers. For amplitudes folk, this is the stuff of dreams. Up until recently the best we could achieve was an expansion in small or large parameters!

So what’s new? Dr. Amit Sever brought us up to date on developments at the Perimeter Institute, where the world’s most brilliant minds have found a way to map certain scattering amplitudes in $4$ dimensions onto a $2$ dimensional model which can be exactly solved. More technically, they’ve created a flux tube representation for planar amplitudes in $\mathcal{N}=4$ super-Yang-Mills, which can then by solved using spin chain methods.

The upshot is that they’ve calculated $6$ particle scattering amplitudes to all values of the (‘t Hooft) coupling. Their method makes no mention of Feynman diagrams or string theory – the old-fashioned ways of computing this amplitude for weak and strong coupling respectively. Nevertheless the answer matches exactly known results in both of these regimes.

There’s more! By putting their computation under the microscope they’ve unearthed unexpected new physics. Surprisingly the multiparticle poles familiar from perturbative quantum field theory disappear. Doing the full calculation smoothes out divergent behaviour in each perturbative term. This is perhaps rather counterintuitive, given that we usually think of higher-loop amplitudes as progressively less well-behaved. It reminds me somewhat of Regge theory, in which the UV behaviour of a tower of higher spin states is much better than that of each one individually.

The smorgasbord of progress continued in Mattias Wilhelm’s talk. The Humboldt group have a completely orthogonal approach linking integrability to amplitudes. By computing form factors using unitarity, they’ve been able to determine loop-corrections to anomalous dimensions. Sounds technical, I know. But don’t get bogged down! I’ll give you the upshot as a headline – New Link between Methods, Form Factors Say.

Coffee consumed, and it was time to get colorful. You’ll hopefully remember that the quarks holding protons and neutrons together come in three different shades. These aren’t really colors that you can see. But they are internal labels attached to the particles which seem vital for our theory to work!

About 30 years ago, people realised you could split off the color-related information and just deal with the complicated issues of particle momentum. Once you’ve sorted that out, you write down your answer as a sum. Each term involves some color stuff and a momentum piece. Schematically

$\displaystyle \textrm{gluon amplitude}=\sum \textrm{color}\times \textrm{kinematics}$

What they didn’t realise was that you can shuffle momentum dependence between terms to force the kinematic parts to satisfy the same equations as the color parts! This observation, made back in 2010 by Zvi Bern, John Joseph Carrasco and Henrik Johansson has important consequences for gravity in particular.

Why’s that? Well, if you arrange your Yang-Mills kinematics in the form suggested by those gentlemen then you get gravity amplitudes for free. Merely strip off the color bit and replace it by another copy of the kinematics! In my super-vague language above

$\displaystyle \textrm{graviton amplitude}=\sum \textrm{kinematics}\times \textrm{kinematics}$

Dr. John Joseph Carrasco himself brought us up to date with a cunning method of determining the relevant kinematic choice at loop level. I can’t help but mention his touching modesty. Even though the whole community refers to the relations by the acronym BCJ, he didn’t do so once!

Before that Dr. Donal O’Connell took us on an intriguing detour of solutions to classical gravity theories with an appropriate dual Yang-Mills theory, obtainable via a BCJ procedure. The idea is beautiful, and seems completely obvious once you’ve been told! Kudos to the authors for thinking of it.

After lunch we enjoyed a well-earned break with a hike up the Uetliberg mountain. I learnt that this large hill is colloquially called Gmuetliberg. Yvonne Geyer helpfully explained that this is derogatory reference to the tame nature of the climb! Nevertheless the scenery was very pleasant, particularly given that we were mere minutes away from the centre of a European city. What I wouldn’t give for an Uetliberg in London!

Evening brought us to Heidi and Tell, a touristic yet tasty burger joint. Eager to offset some of my voracious calorie consumption I took a turn around the Altstadt. If you’re ever in Zurich it’s well worth a look – very little beats medieval streets, Alpine water and live swing music in the evening light.

Conversations

It was fantastic to meet Professor Lionel Mason and discuss various ideas for extending the ambitwistor string formalism to form factors. I also had great fun chatting to Julio Martinez about linking CHY and BCJ. Finally huge thanks to Dr. Angnis Schmidt-May for patiently explaining the latest research in the field of massive gravity. The story is truly fascinating, and could well be a good candidate for a tractable quantum gravity model!

Erratum: An earlier version of this post mistakenly claimed that Chris White spoke about BCJ for equations of motion. Of course, it was his collaborator Donal O’Connell who delivered the talk. Many thanks to JJ Carrasco for pointing out my error!

# Conference Amplitudes 2015 – Integration Ahoy!

I recall fondly a maths lesson from my teenage years. Dr. Mike Wade – responsible as much an anyone for my scientific passion – was introducing elementary concepts of differentiation and integration. Differentiation is easy, he proclaimed. But integration is a tricky beast.

That prescient warning perhaps foreshadowed my entry into the field of amplitudes. For indeed integration is of fundamental importance in determining the outcome of scattering events. To compute precise “loop corrections” necessarily requires integration. And this is typically a hard task.

Today we were presented with a smorsgasbord of integrals. Polylogarithms were the catch of the day. This broad class of functions covers pretty much everything you can get when computing amplitudes (provided your definition is generous)! So what are they? It fell to Dr. Erik Panzer to remind us.

Laymen will remember logarithms from school. These magic quantities turn multiplication into addition, giving rise to the ubiquitous schoolroom slide rules predating electronic calculators. Depending on your memory of math class, logarithms are either curious and fascinating or strange and terrifying! But boring they most certainly aren’t.

One of the most amusing properties of a logarithm comes about from (you guessed it) integration. Integrating $x^{a-1}$ is easy, you might recall. You’ll end up with $x^a/a$ plus some constant. But what happens when $a$ is zero? Then the formula makes no sense, because dividing by zero simply isn’t allowed.

And here’s where the logarithm comes to the rescue. As if by witchcraft it turns out that

$\displaystyle \int_0^x x^{-1} = -\log (1-x)$

This kind of integral crops when you compute scattering amplitudes. The traditional way to work out an amplitudes is to draw Feynman diagrams – effectively pictures representing the answer. Every time you get a loop in the picture, you get an integration. Every time a particle propagates from A to B you get a fraction. Plug through the maths and you sometimes see integrals that give you logarithms!

But logarithms aren’t the end of the story. When you’ve got many loop integrations involved, and perhaps many propagators too, things can get messy. And this is where polylogarithms come in. They’ve got an integral form like logarithms, only instead of one integration there are many!

$\displaystyle \textrm{Li}_{\sigma_1,\dots \sigma_n}(x) = \int_0^z \frac{1}{z_1- \sigma_1}\int_0^{z_1} \frac{1}{z_2-\sigma_2} \dots \int_0^{z_{n-1}}\frac{1}{z_n-\sigma_n}$

It’s easy to check that out beloved $\log$ function emerges from setting $n=1$ and $\sigma_1=0$. There’s some interesting sociology underlying polylogs. The polylogs I’ve defined are variously known as hyperlogs, generalized polylogs and Goncharov polylogs depending on who you ask. This confusion stems from the fact that these functions have been studied in several fields besides amplitudes, and predictably nobody can agree on a name! One name that is universally accepted is classical polylogs – these simpler functions emerging when you set all the $\sigma$s to zero.

So far we’ve just given names to some integrals we might find in amplitudes. But this is only the beginning. It turns out there are numerous interesting relations between different polylogs, which can be encoded by clever mathematical tools going by esoteric names – cluster algebras, motives and the symbol to name but a few. Erik warmed us up on some of these topics, while also mentioning that even generalized polylogs aren’t the whole story! Sometimes you need even wackier functions like elliptic polylogs.

All this gets rather technical quite quickly. In fact, complicated functions and swathes of algebra are a sad corollary of the traditional Feynman diagram approach to amplitudes. But thankfully there are new and powerful methods on the market. We heard about these so-called bootstraps from Dr. James Drummond and Dr. Matt von Hippel.

The term bootstrap is an old one, emerging in the 1960s to describe methods which use symmetry, locality and unitarity to determine amplitudes. It’s probably a humorous reference to the old English saying “pull yourself up by your bootstraps” to emphasise the achievement of lofty goals from meagre beginnings. Research efforts in the 60s had limited success, but the modern bootstrap programme is going from strength to strength. This is due in part to our much improved understanding of polylogarithms and their underlying mathematical structure.

The philosophy goes something like this. Assume that your answer can be written as a polylog (more precisely as a sum of polylogs, with the integrand expressed as $\prod latex d \log(R_i)$ for appropriate rational functions $R_i$). Now write down all the possible rational functions that could appear, based on your knowledge of the process. Treat these as alphabet bricks. Now put your alphabet bricks together in every way that seems sensible.

The reason the method works is that there’s only one way to make a meaningful “word” out of your alphabet bricks. Locality forces the first letter to be a kinematic invariant, or else your answer would have branch cuts which don’t correspond to physical particles. Take it from me, that isn’t allowed! Supersymmetry cuts down the possibilities for the final letter. A cluster algebra ansatz also helps keep the possibilities down, though a physical interpretation for this is as yet unknown. For $7$ particles this is more-or-less enough to get you the final answer. But weirdly $6$ particles is smore complicated! Counter-intuitive, but hey – that’s research. To fix the six point result you must appeal to impressive all-loop results from integrability.

Next up for these bootstrap folk is higher loops. According to Matt, the $5$-loop result should be gettable. But beyond that the sheer number of functions involved might mean the method crashes. Naively one might expect that the problem lies with having insufficiently many constraints. But apparently the real issue is more prosaic – we just don’t have the computing power to whittle down the options beyond 5-loop.

With the afternoon came a return to Feynman diagrams, but with a twist. Professor Johannes Henn talked us through an ingenious evaluation method based on differential equations. The basic concept has been known for a long time, but relies heavily on choosing the correct basis of integrals for the diagram under consideration. Johannes’ great insight was to use conjectures about the dlog form of integrands to suggest a particularly nice set of basis integrals. This makes solving the differential equations a cinch – a significant achievement!

Now the big question is – when can this new method be applied? As far as I’m aware there’s no proof that this nice integral basis always exists. But it seems that it’s there for enough cases to be useful! The day closed with some experimentally relevant applications, the acid test. I’m now curious as to whether you can link the developments in symbology and cluster algebras with this differential equation technique to provide a mega-powerful amplitude machine…! And that’s where I ought to head to bed, before you readers start to worry about theoretical physicists taking over the world.

Conversations

It was a pleasure to chat all things form factors with Brenda Penante, Mattias Wilhelm and Dhritiman Nandan at lunchtime. Look out for a “on-shell” blog post soon.

I must also thank Lorenzo Magnea for an enlightening discussion on soft theorems. Time to bury my head in some old papers I’d previously overlooked!

# Conference Amplitudes 2015!

It’s conference season! I’m hanging out in very warm Zurich with the biggest names in my field – scattering amplitudes. Sure it’s good fun to be outside the office. But there’s serious work going on too! Research conferences are a vital forum for the exchange of ideas. Inspiration and collaboration flow far more easily in person than via email or telephone. I’ll be blogging the highlights throughout the week.

Monday | Morning Session

To kick-off we have some real physics from the Large Hadron Collider! Professor Nigel Glover‘s research provides a vital bridge between theory and experiment. Most physicists in this room are almost mathematicians, focussed on developing techniques rather than computing realistic quantities. Yet the motivation for this quest lie with serious experiments, like the LHC.

We’re currently entering an era where the theoretical uncertainty trumps experimental error. With the latest upgrade at CERN, particle smashers will reach unprecedented accuracy. This leaves us amplitudes theorists with a large task. In fact, the experimentalists regularly draw up a wishlist to keep us honest! According to Nigel, the challenge is to make our predictions twice as good within ten years.

At first glance, this 2x challenge doesn’t seem too hard! After all Moore’s Law guarantees us a doubling of computing power in the next few years. But the scale of the problem is so large that more computing power won’t solve it! We need new techniques to get to NNLO – that is, corrections that are multiplied by $\alpha_s^2$ the square of the strong coupling. (Of course, we must also take into account electroweak effects but we’ll concentrate on the strong force for now).

Nigel helpfully broke down the problem into three components. Firstly we must compute the missing higher order terms in the amplitude. The start of the art is lacking at present! Next we need better control of our input parameters. Finally we need to improve our model of how protons break apart when you smash them together in beams.

My research helps in a small part with the final problem. At present I’m finishing up a paper on subleading soft loop corrections, revealing some new structure and developing a couple of new ideas. The hope is that one day someone will use this to better eliminate some irritating low energy effects which can spoil the theoretical prediction.

In May, I was lucky enough to meet Bell Labs president Dr. Marcus Weldon in Murray Hill, New Jersey. He spoke about his vision for a 10x leap forward in every one of their technologies within a decade. This kind of game changing goal requires lateral thinking and truly new ideas.

We face exactly the same challenge in the world of scattering amplitudes. The fact that we’re aiming for only a 2x improvement is by no means a lack of ambition. Rather it underlines that problem that doubling our predictive power entails far more than a 10x increase in complexity of calculations using current techniques.

I’ve talked a lot about accuracy so far, but notice that I haven’t mentioned precision. Nigel was at pains to distinguish the two, courtesy of this amusing cartoon.

Why is this so important? Well, many people believe that NNLO calculations will reduce the renormalization scale uncertainty in theoretical predictions. This is a big plus point! Many checks on known NNLO results (such as W boson production processes) confirm this hunch. This means the predictions are much more precise. But it doesn’t guarantee accuracy!

To hit the bullseye there’s still much work to be done. This week we’ll be sharpening our mathematical tools, ready to do battle with the complexities of the universe. And with that in mind – it’s time to get back to the next seminar. Stay tuned for further updates!

Update | Monday Evening

Only time for the briefest of bulletins, following a productive and enjoyable evening on the roof of the ETH main building. Fantastic to chat again to Tomek Lukowski (on ambitwistor strings), Scott Davies (on supergravity 4-loop calculations and soft theorems) and Philipp Haehnal (on the twistor approach to conformal gravity). Equally enlightening to meet many others, not least our gracious hosts from ETH Zurich.

My favourite moment of the day came in Xuan Chen’s seminar, where he discussed a simple yet powerful method to check the numerical stability of precision QCD calculations. It’s well known that these should factorize in appropriate kinematic regions, well described by imaginatively named antenna functions. By painstakingly verifying this factorization in a number of cases Xuan detected and remedied an important inaccuracy in a Higgs to 4 jet result.

Of course it was a pleasure to hear my second supervisor, Professor Gabriele Travaglini speak about his latest papers on the dilatation operator. The rederivation of known integrability results using amplitudes opens up an enticing new avenue for those intrepid explorers who yearn to solve $\mathcal{N}=4$ super-Yang-Mills!

Finally Dr. Simon Badger‘s update on the Edinburgh group’s work was intriguing. One challenge for NNLO computations is to understand 2-loop corrections in QCD. The team have taken an important step towards this by analysing 5-point scattering of right-handed particles. In principle this is a deterministic procedure: draw some pictures and compute.

But to get a compact formula requires some ingenuity. First you need appropriate integral reduction to identify appropriate master integrals. Then you must apply KK and BCJ relations to weed out the dead wood that’s cluttering up the formula unnecessarily. Trouble is, both of these procedures aren’t uniquely defined – so intelligent guesswork is the order of the day!

That’s quite enough for now – time for some sleep in the balmy temperatures of central Europe.

# Science and Faith – The Arts of the Unknown

I spent this morning singing a Sunday service at St. George’s Church in Borough. An odd occupation for a scientist perhaps, especially given the high profile of several atheist researchers! Yet a large number of scientists see no contradiction between faith and science. In fact, my Christian faith is only deepened by my fascination with the natural world.

Picture a scientist. Chances are you’ve already got in your mind a geeky, rational person, calibrating a precise experiment or poring over a dry mathematical formula! As with any stereotype, it has it’s merits. But it misses a vital quality in research – imagination.

To succeed as a scientist, you must be creative above all else. It’s no use just learning experimental techniques or memorising formulae. Every new idea must necessarily start off as a fantasy. Great painters are not merely lauded for their 10,000 hours of practice with a paintbrush. It is their capacity to conceive and relay vivid scenes which ensures their place in history. And so it is with science.

So why are scientists seen as cold and calculating and exact, rather than passionate and original? The problem lies in education. While young children are encouraged to express themselves in Literacy, Numeracy is all too often a trudge through tedious and predictable sums. In “arts” subjects, questions are a magical tool enabling discussion, debate and opinion. In “sciences” they merely distinguish right from wrong.

After 15 years of schooling, no wonder the stereotype is embedded! As a teenager, I very nearly ditched the sciences in favour of subjects where expression was free and original arguments rewarded. I’m eternally thankful to my teachers, parents and bookshelf for convincing me that the curriculum was utterly unrepresentative of real science.

So what’s to be done. For any budding scientists out there, your best bet is to read some books. Not your school textbooks – chances are they are dull as ditchwater and require no creative input at all. I mean books written by real life mathematicians, physicists, biologists… These will give you an insight into the imagination that drives research, the contentious debates and the lively exchanges of ideas.

You might not understand everything, but that’s the whole point – science is about the unknown, just as much as art or faith. It is exactly this point which we must evangelise again and again. Perhaps then fewer people will write negative reviews criticising science for being complex, poetic and beautiful.

As a wider society, we can take action too! We must demand better science teaching from a young age. Curricula should emphasise problem solving over knowledge, ideas over techniques and originality over regurgitation. This is already the mantra for many traditionally “artistic” discplines. It must be the battle cry for scientists also!

A better approach to science would democratize opportunity for the next generation. No longer will the relative creativity of girls be arbitrarily punished – an approach which can only discourage women from entering science in the long run. No longer will there be a tech skills gap threatening to undermine the thriving software industry. The UK has a uniquely privileged scientific pedigree. For future equality, economy and diversity, we must use it.

In the service this morning Fr Jonathan Sedgwick talked of the danger of applying scientific laws to the world at large. The concepts of “cause and effect” and “zero sum games” may well work in vacuo, but they are artificial and burdensome when applied to interpersonal relationships. Quite right – as Christians we must question these human rules, and look for a divine inspiration to guide our lives!

But this is also precisely what we must do as scientists. A good scientist always questions their models, constantly listening for the voice of intuition. For science – like our own existence – is ever changing. And it’s our job to search for the way, the truth and the life.

My thanks to Margaret Widdess, who prepared me for confirmation two years ago at St. Catharine’s College, Cambridge and with whom I first talked deeply about the infinity of science and faith.

# Why does feedback hurt sometimes?

Research is hard. And not for the reasons you might expect! Sure, my daily life involves equations which look impenetrable to the layman. But by the time you’ve spent years studying them, they aren’t so terrifying!

The real difficulty in research is psychological. The natural state for a scientist is failure – most ideas simply do not succeed! Developing the resilience, maturity and sheer bloody mindedness to just keep on plugging away is a vital but tough skill.

This letter, written by an experienced academic to her PhD student is a wonderfully candid account of the minefield of academic criticism, both professional and personal. What’s more, it lays bare some important coping strategies – I certainly wish I’d read it before embarking on my PhD.

Above all, this letter is an admission of humanity. As researchers, we face huge challenges in our careers. But the very personal process of responding to them is precisely what makes us better scientists, and perhaps even improves us as people.

This letter was written by an experienced academic at ANU to her PhD student, who had just presented his research to a review panel and was still licking her wounds.

The student sent it to me and I thought it was a great response I asked the academic in question, and the student who received it, if I could publish it. I wish all of us could have such nuanced and thoughtfu feedback during the PhD. I hope you enjoy it.

A letter to…My PhD student after her upgradeWell you did it. You got your upgrade. But from the look on your face I could tell you thought it was a hollow victory. The professors did their job and put the boot in. I remember seeing that look in the mirror after my own viva. Why does a win in academia always have the sting of defeat?

Yeah, it’s a…

View original post 1,120 more words

## Quantum or Otherwise

the morning paper

a random walk through Computer Science research, by Adrian Colyer

Gödel's Lost Letter and P=NP

a personal view of the theory of computation

Seeing Things

Why we see what we see

Sci-5

Five minutes with a scientist

The Thesis Whisperer

Just like the horse whisperer - but with more pages

Of Particular Significance

Conversations About Science with Theoretical Physicist Matt Strassler

Gowers's Weblog

Mathematics related discussions

4 gravitons

The trials and tribulations of four gravitons and a postdoc

Eventually Almost Everywhere