Hacker Newsnew | past | comments | ask | show | jobs | submitlogin
The Deconstructed Standard Model Equation (symmetrymagazine.org)
86 points by bit-player on July 29, 2016 | hide | past | favorite | 54 comments


I've been banging my head on QFT and the Standard Model for enough time now that I can actually glimpse more or less what those terms actualy mean, as in , I can translate them into words. It's almost like learning a new language! Pretty awesome.

Still, despite all this complexity, we still have several minor and major theoretical gaps to fill. Exciting times!


I, as a layman with basic college physics knowledge, and passing knowledge of the standard model, want to understand the meaning of those terms. Do you know how I can go about that? Ideally, with something a little less dense than a textbook.


I have a PhD in physics, though not particle physics. QFT was required, since the methods appear all over in condensed matter physics.

Unfortunately, there is no real shortcut to being able to understand those terms except by studying one of the standard texts. And, physics being what it is, you can expect a pretty hard slog because the texts will assume you know first quantization back and forth. Add to that, it's not just QFT but the domain-specific standard model knowledge.

Dealing with QM requires PDEs, linear algebra, a fair bit of applied analysis (real and complex). Dealing with QFT requires that, plus learning a bunch of new techniques with the typical hand-wavy rigor of physics. (Hand-wavy compared to math.)

If you want to get a feel of the Standard Model without knowing QFT, the text I used when I took senior-level particle physics is great: Introduction to Elementa Particles, by David Griffiths. (He's an excellent writer, btw; I recommend any of his physics books).

As for the standard texts for particle physics, I'm afraid I've been out of the loop for too long to remember what books were used.


A related question: Feynman's popular book QED claims to explain quantum electrodynamics enough that you could almost do calculations with it, just ridiculously inefficiently. After reading it, I can't: the details left out, I can't easily fill in from the grad-level texts, even with a pretty decent undergrad physics background. Shouldn't it be possible to explain QED to a programmer using a literate program? (Again, an inefficient one.) Was Feynman exaggerating? Or is it just the tininess of the market of programmers who want to understand what QED is really about who aren't out to become professional physicists?

(For others, QED is the part of the standard model about electrons and light -- the most relevant part for everyday physics and apparently the simplest part too.)


Apart the character of physical laws and the easy pieces, what's the gateway book to understand feynann more advanced work?


I don't really know, because I don't understand the more advanced work like QED. But http://www.feynmanlectures.caltech.edu/I_toc.html is one of my favorite books ever.


If you are looking for the book, the name is "Introduction to Elementary Particles" (not Elementa which is a typo).


I was going to recommend the Griffiths book as well. The first two chapters are readable without any demanding mathematics. The latter ones require more work, but build up nicely.


Zee's QFT in a Nutshell is a very readable, high-level view of what's going on in QFT. Griffiths' book is his one text I haven't read, but I love his others so much I can't help but second the recommendation.


I'm afraid there really isn't another way around. This stuff is just very difficult. If you want to understand the standard model you need to understand Quantum Field Theory, and to understand Quantum Field Theory you need to understand Quantum Mechanics, and to understand that you need to understand classical hamiltonian and lagrangian mechanics, etc etc, and to understand all that you need lots of mathematical notions, so... There really isn't a fast way but to study the standard texts.


Can you recommend how to get started on this? Like beginner links or something? I'm currently just a physics noob, but have some of the basic concepts of quantum theory down. I would love to just have a basic grasp of that monstrous equation.


"Quantum Field Theory for the Gifted Amateur" by Stephen Blundell and Tom Lancaster.

It might be right for you if you know the basics of Quantum Mechanics and electromagnetism


I would love to see this article, about 5-10 times longer. Dig one level deeper into what the notation means, how it relates to the zoo of fundamental particles, etc. Compare and contrast the sections dealing with different forces/particles. Talk something about the history of how this came out as a unification of models for gravity, electricity and magnetism, weak force, strong force.


I love how everybody uses Gutierrez's typesetting. If I ever meet him and have the presence of mind to put two and two together, I'll buy him a beer.

If you want the TeX source for your own purposes, it's here: http://nuclear.ucdavis.edu/~tgutierr/files/stmL1.html


What is the sum of terms supposed to be equal to?


It's a Lagrangian, so this is the function nature tries to minimize. In common everyday systems, the Lagrangian is kinetic energy minus potential energy.


True, but with nitpick that it is the Lagrangian density so nature minimises its total over a region of 4-space.

And with the more important nitpick that this Lagrangian is a quantum operator rather than a number. In some sense nature does try to minimise it even so, but I never got an intuitive grasp on what that sense is.


And with the more important nitpick that this Lagrangian is a quantum operator rather than a number.

I don't think that's true when dealing with the formulation in terms of path integrals as we do here...


It's the focus on path integrals that obscures the fact that the Lagrangian is an operator. Except maybe for convenience, there is nothing about such equations that is unique to the path-integral formulation.

The fields, the Langragian, and the scattering matrix are all operators; but of course their matrix elements are numbers. The path-integral formulation is a good -- and I think physically well motivated -- trick for calculating those matrix elements in terms of merely fields and their associated Largrangians.


The text however refers to it as an equation, and that's not an equation.


Anything. It's a state function measuring energy, so pick a number and keep it the same. Usually zero is a good choice.


"To clean up these redundancies, theorists use virtual particles they call ghosts.

This part of the equation describes how matter particles interact with Higgs ghosts, virtual artifacts from the Higgs field. [...] This last part of the equation includes more ghosts. These ones are called Faddeev-Popov ghosts, and they cancel out redundancies that occur in interactions through the weak force."

So the second half this equation is used to describe invisible things needed to cancel out wrong stuff from the first half? Sounds ad hoc to me, were these "ghosts" predicted by anyone beforehand? Even if not, as a model it can still be useful though.


Faddeev–Popov ghosts in particular are artifacts of covariant quantization in the path integral formalism. One way to think of them is as a kind of Jacobian determinant of the transformation to physical degrees of freedom.


The two types of "ghosts" here are very different. In both cases, though, they're mathematical artefacts rather than anything "physical", but I'll try to explain them as well as I know. Disclaimer: The most advanced physics I've done was a first course in this stuff, so I might be wrong about some things.

For what they're calling "Higgs ghosts": There are two different ways of describing the electromagnetic and weak forces, and which one is best depends on how much energy the particles you're dealing with have. At really high energies, it makes the most sense to talk about a combined "electroweak force", described in terms of four fields with 2 components each (often called W1, W2, W3 and B), and one 4-component field (the Higgs field).

In contrast, at low energies, it makes more sense to talk about the electromagnetic force, with one 2-component field (y), and the weak force, with three 3-component fields (W+, W-, Z0) and a 1-component field (the Higgs field, again). So, where did the other three components of the Higgs field go? Well, we just rearranged things - if you check, the total number of components stayed the same. There are various names for this rearrangement, and I haven't seen this one before, but I guess they're calling these "missing" components "ghosts".

As for the other type, the Faddeev-Popov ghosts, those are more obviously mathematical artefacts. Normally, you'd start by writing down a "physical" Lagrangian ("physical" here meaning something like "written in terms of actual physical fields").

But it turns out that you can't actually calculate with the physical Lagrangian. So you have to rewrite it in a (mostly) mathematically-equivalent way, which involves extra fields. These fields come along with extra rules which basically say "no state you can actually measure involves the ghost fields in any way". Really, they're just there as a calculational aid and aren't physically "real", and they're called "ghosts" to reflect that.

Hope that's at least vaguely comprehensible, it's difficult to explain this stuff without assuming a lot of background knowledge.


Mostly right, but one correction:

You actually _can_ calculate with the physical Lagrangian. This is what lattice gauge theory simulations do. But it's inconvenient and difficult in perturbation theory, so physicists use the Fadeev-Popov ghost trick instead. The resulting computations are _entirely_ mathematically equivalent, not "(mostly)".


It's the same as noting you can more quickly calculate 19x21 by noting it's equivalent to 20x20 - 1. This ghostly term is the -1


It doesn't really make sense to ask if the ghosts were predicted. They're mathematical artifacts of a particular computational formalism. There are other computational formalisms for QFT which don't feature ghosts, and which give the same answers for any observable quantities.


"Note: ...In Gutierrez’s dissemination of the transcript, he noted a sign error he made somewhere in the equation. Good luck finding it!"

Anyone see the flipped sign?


Given the world we live in, right now, why does mathematics continue to insist on minified expressions?

Given the option, most development teams would choose to read and write against verbose source code, rather than scrape obfuscated variables and method signatures out of a minified, transpiled, compressed package.

So why do we continue this archaic practice of obscure, inscrutable symbols in mathematics? Cultural inertia?

The cycle of madness must end!


Your comment shows a common misunderstanding of what mathematicians are trying to do. Modern mathematical notation is not obfuscated, it is in fact making the object being described much easier to perceive for humans.

Explaining the equation you see in English words is what every physics book does, however the equation itself represents a concept that is not human. It comes from an alien universe of symmetries and relations and we have spent centuries to arrive at the current way of writing these down in a way which makes them easy to work with.

You can formulate the solution to a quadratic equation as:

The negative linear term, added and subtracted from the square root of the quantity which is the difference between the square of the linear term and the product of four times the constant term and the quadratic term, all divided by twice the quadratic term.

Modern mathematicians write:

x = (-b +- sqrt(b^2 - 4ac)) / 2a

The difference is night and day. You cannot remove the essential complexity from a problem. You can only try to get close to its representation.


While this might not be the OP's point, but for me the main problem are the one-character variable names. Generally computer code that just operates on cryptic one-letter variable names, is considered bad code. Yet equations almost always have one letter variables and the worst part of all, their meaning is almost never explained anywhere

Case in point, what are b, a and c? what is their meaning if I have to measure them, what is the device for measuring them or what quantity do they represent?


The point is that variables are evil and arbitrary. So you give them short names to signal that you could replace them with anything else, what matters is how they appear together. Additionally different subfields of mathematics and physics usually adopt naming conventions.


Took me years before thinking of variables that way and mathematical expressions are constraints over sets, formulas expressing relations of interests that way.


>Case in point, what are b, a and c? what is their _meaning_

The "b, a, and c" in the quadratic equation are abstractions. They are placeholders for any number. That math equation is purposely designed to not have any concrete meaning for a,b,c. Therefore, asking for their meaning does't make any sense.

I'll try an analogy. We can give provide analogous "computer equation" for taking the last four characters of a string:

  s2 = Right(s1, 4);
If someone were to ask "what is the _meaning_ of s1?", it would be a nonsensical question. The s1 is a placeholder for any string. The s1 could be a domain name and you want to extract the TLD (".com", ".org", ".edu"). Or the s1 could be a date "mm-dd-yyyy" and you want to extract the "2016" from ""10-31-2016". The s1 could be a placeholder for a thousand unrelated applications.

If you further generalize the "computer equation" to be take the rightmost "x" amount of characters, you'd replace the "4" with "x" like this:

  s2 = Right(s1, x);
... it could trigger confusion as to "what does x _mean_? And what measuring device does it come from?"

The "x" is an abstraction/parameter/argument of any number. That number could be 4, or 40000. (You might be tempted to say, "but I know 'x' stands for the number of characters I want from the string!" If so, by analogy you can also say that "c" in the quadratic equation stands for the number that satisfies the equation when you know "x", "a", and "b" ... which just circles back to each 1-character variable not having a concrete "meaning."

The concrete meaning of "a,b,c" in the quadratic equation depends on whether the equation was modeling something in physics, or engineering, or finance.[1][2]

The concrete meaning of "s1" and "x" in the "computer equation" depends on whether you're parsing logs, decomposing dates, verifying credit card numbers, etc.

The analogy above could have been demonstrated with regex such as "(.+)(.{4})" ... which would trigger questions, *"why do programmers use 1-character symbols such as '.' and '+' instead of "anycharacterm_oneormoretimes_consideredasagroup_then_anycharacters_fourtimes_consideredasagroup" ... etc

tldr: computer programmers also deal with weird-looking "equations" but since they are used to them, they have the background knowledge to decipher them.

[1]https://www.mathsisfun.com/algebra/quadratic-equation-real-w...

[2]https://www.boundless.com/algebra/textbooks/boundless-algebr...


Thank you for the thorough explanation!

I see your point with abstract expressions, but the connection I'm failing to make is - if they are completely abstract, then how can the equation be related to the physical world?

In essence, the equation defines the relations between some values - but if any or some of the values can be replaced by real measurements, then surely the relations between real measurements are not interchangeable?


if they are completely abstract, then how can the equation be related to the physical world?

Math equations do not have to be related to anything in the physical world. It's just that humans have found that some patterns in nature seem to match some equations. Why or how math describes the real world is a deep philosophical question![1]

When G.H. Hardy[2] was exploring certain aspects of number theory and properties of prime numbers, it didn't have any relevance to the real world. 100 years later, we now use the math concepts in public key cryptography.

Does Isaac Newton's "F = ma" of "force = mass times acceleration" describe the "reality" of gravity and how planets move in the solar system? Well -- it does -- and it doesn't. It seemed to match reality well enough to find the planet Neptune -- purely by using math. But Newton's math equations didn't seem to match the "reality" of Mercury's orbit.

Then Einstein came up with another set of math equations that matched the "reality" of Mercury's shifting orbit.[3] But now today's physicists notice that Einstein's formulas don't match the "reality" of galaxies rotating faster than the his formulas predict. Today's physicists are trying to come up with another set of math equations to match that reality (or find the elusive "dark matter" so that Einstein's equations remain unchanged.)

(Einstein's equations do not make Newtons's equations obsolete. Newton's math (that doesn't perfectly match reality) is still good enough to guide Apollo rockets to the Moon and back. It's just not good enough to fully describe how Mercury wiggles around the Sun.)

Math is not reality. Math is its own topic that can stand apart from "physics" or "engineering". (Hence, math is the ultimate abstraction.) It's just sometimes convenient for humans to map some mathematics to some realities. It's often helpful for us to do that. Sometimes it backfires. Some Wall Street guys "mapped" math equations to model the financial behavior of mortgages and they turned out to be wrong.

Put another way: If math is used to model the physical world, it is only an approximation for it.

[1]https://www.google.com/search?q=the+unreasonable+effectivene...

[2]https://en.wikipedia.org/wiki/A_Mathematician%27s_Apology#Cr...

[3]http://physics.stackexchange.com/questions/26408/what-did-ge...


> Math equations do not have to be related to anything in the physical world.

Not to derail this into philosophy, but that position is conjecture. There are people who adopt a position where all mathematical concepts actually derive from physical structures, ie. a kind of mathematical physicalism. These people hold that the reduction of math to the physical may be obscure, like a reduction of economics to physics, but still real.


Though I can't speak for the parent post's author, applying that post's spirit to your example would instead yield, for the quadratic formula:

'The negative linear term, added and subtracted from the square root of the discriminant, all divided by twice the quadratic term'; then before or after this you'd define the discriminant, as it's an illuminating concept in its own right.

I see the target of this wishful thinking a sort-of middle-ground between a formula's derivation/proof and its compact form.


This is rather poetic. There is no difference, mathematical notation is a natural language and initiated mathematicians will read it in sentence form.

Reading is a process of form recognition disconnect from vocalization. On the other hand, language seems to be mostly perceived as vocalisation, because that's the form it's learned from. As anecdotal evidence, tongue is an idiom for language and people have to make an actual effort to read without vocalizatio in order to increase reading-speed.

Note that your literal example does contain a lot more information about context. IE, you even omitted the quadratic equation itself. That shows exactly the problem that I thought the Parent would mean:

I prefer source code in a given language, that's actually compiled and maybe even proven correct, over terse mathematical notation that leaves a lot to context, intuition and shared information. Computability Theory gave rise to Computers, not the other way around.


Your assessment is somewhat unfair and kind of misguided. The differences between math and programming are quite large. I think there are two main reasons for why we have symbols in math and more verbose expressions in programming.

1. Medium. Math is mainly done with pen/pencil on paper, or chalk on a blackboard, or marker on a whiteboard. Programming is mainly done with keyboard on a computer. Those media are quite different. In a way, it's about ease of input - it's easier to write out a longer name with keyboard than with a pen.

2. Nature of the activity. A lot of math boils down to manipulating expressions given rules. It involves a lot of scratch work with a lot of backtracking and trying new things. A lot of programming boils down to easily reading and writing code and reasoning about how things work. Imagine writing "integral from a to b of f(x) dx" or "integral(f(x), x, a, b)" by hand a hundred times vs in an IDE and you'll see what I mean.


That's not how math statements work.

You think mathematicians are minifying/uglifying their 'code' in the same way that one might use switch/case or ternary expressions instead of if/else if/else statements: in the name of concision/premature optimisation.

However, what the mathematical language is really doing is more akin to what jQuery is doing for JavaScript: it wraps low-level, complex and at times verbose mathematical statements into simpler representations.

Minification !== abstraction/syntactic sugar.


A lot of mathematical notation is optimized for hand-written algebra.

One of the main things you "do" with expressions like these is to transform them into equivalent expressions that have some desirable form, or examine special cases that allow simplifications, or examine the result of applying some additional operation.

Compact symbols and single letter variables reduce the amount of writing you need to do for each step of such transformations, but they also make the "structure" of expressions more visually obvious.

Programmers also spend time transforming programs into equivalent programs (aka refactoring), but refactoring isn't quite as important to programmers as algebra is to mathematicians and physicists, so (popular) programming languages don't optimize for it to quite the same degree.


>Given the world we live in, right now, why does mathematics continue to insist on minified expressions?

Because it is concise, precise, and widely accepted.

>Given the option, most development teams would choose to read and write against verbose source code, rather than scrape obfuscated variables and method signatures out of a minified, transpiled, compressed package.

What's your point? Source code is not the same thing as mathematics, not for the majority of software and not for the majority of mathematics as practiced by mathematicians. Verbosity makes sense when your domain involves concrete entities like "customers" and "widgets" and "thermal sensors." When your domain involves abstract entities like "ring homomorphisms" and "clopen sets" and "vector spaces," it doesn't. If you're writing a database, a name like "transaction_mutex" is more descriptive than "m." If you're universally quantifying over the domain of an arbitrary continuous function on the reals, "x" is about as descriptive -- and as conventional -- as any name you can come up with.

As an aside, I really wish we would dispense with the non-word "transpile." We have a word for translating a program from one language to another: compile.

>So why do we continue this archaic practice of obscure, inscrutable symbols in mathematics?

Because it works very well. I'm not sure what else to say. You spend maybe fifteen minutes learning about, say, the symbol "∂" when you're introduced to multivariate calculus, and then for the rest of time you have an extremely concise way of expressing a variety of combinations of partial derivatives that can be understood by anyone who has also been introduced to multivariate calculus.

Don't get me wrong, there are actual problems with mathematical notation -- overloading, "abuse of notation," and as often as not just plain omitting information -- but the use of non-ASCII symbols and short variable names are not among them.


Am I correct in understanding that this takes every kind of interaction that particles can have, and simply adds them all together?

How does it deal with some particles only interacting in certain ways and not others? Is the user of the equation supposed to make sure to enter zero for those terms, or does the equation capture that knowledge as well?


If you apply an operator of such an interaction to the Lagrangian the equations of motions automatically return zero (e.g. similar to a derivation of a constant).


I haven't seen an article so math intensive in a while... since I saw this one: http://www.scholarpedia.org/article/Bayesian_Ying_Yang_learn...


The article displays a big equation somewhat carefully. But it doesn't have any maths.


If I was a physics phd working on QFT, I would tattoo this equation as a sleeve. Unfortunately, as a mere programmer, I could never pick a sufficiently perfect code snippet to use. I would constantly look at it and try to refactor it.



Nice Y combinator lol. Still more math than code.


There must be some smarter notation that captures the model with less verbosity?

Also, how on earth did various people put this together? Is there a good book about it?


There is; the equation on the CERN mug in the photo is an example. For example, the gluon Lagrangian (Section 1) may be more compactly written

-1/2 Tr F_{\mu\nu} F^{\mu\nu}

for a suitably defined F. Or further abbreviated to something like L_g.

It just depends on how much detail you need to expose.


Well, when you put it like that... suddenly it all makes sense :p

Seriously though, very neat to see the actual equation used to model the universe.


When I was a wee undergrad, a grad student office-mate had printed out these equations and posted them on the office door. He wrote on it: The simplicity of the Standard Model. Or words to that effect; this was in the condensed matter physics division.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: