# Wavefunctions, perspectives, reference frames, representations and symmetries

Ouff ! This title is quite a mouthful, isn’t it? š So… What’s the topic of the day? Well… In our previous posts, we developed a few key ideas in regard to a possible physical interpretation of the (elementary) wavefunction. It’s been an interesting excursion, and I summarized it in another pre-publication paper on the open arXiv.org site.

In my humble view, one of the toughest issues to deal with when thinking about geometric (orĀ physical) interpretations of the wavefunction is the fact that a wavefunction does not seem to obey the classical 360Ā° symmetry in space. In this post, I want to muse a bit about this and show that… Well… It does and it doesn’t. It’s got to do with what happens when you change from one representational base (orĀ representation, tout court)Ā to another which is… Well… Like changing the reference frame but, at the same time, it is also more than just a change of the reference frameāand so that explains the weird stuff (like that 720Ā° symmetry of the amplitudes for spin-1/2 particles, for example).

I should warn you before you start reading: I’ll basically just pick up some statements from my paper (and previous posts) and develop some more thoughts on them. As a result, this post may not be very well structured. Hence, you may want to read the mentioned paperĀ first.

### The reality of directions

Huh? TheĀ realityĀ of directions? Yes. I warned you. This post may cause brain damage. šĀ The whole argument revolves around a thoughtĀ experimentābut one whose results have been verified in zillions of experiments in university student labs so… Well… We do notĀ doubt the results and, therefore, we do not doubt the basic mathematical results: we just want to try to understandĀ them better.

So what is the set-up? Well… In the illustration below (Feynman, III, 6-3), Feynman compares the physics of two situations involving rather special beam splitters. Feynman calls them modified or āimprovedā Stern-Gerlach apparatuses. The apparatus basically splits and then re-combines the two new beams along theĀ z-axis. It is also possible to block one of the beams, so we filter out only particles with their spinĀ upĀ or, alternatively, with their spinĀ down. Spin (or angular momentum or the magnetic moment) as measured along theĀ z-axis, of courseāI should immediately add: we’re talking theĀ z-axis of the apparatus here.

The two situations involve a different relative orientation of the apparatuses: in (a), the angle is 0Ā°, while in (b) we have a (right-handed) rotation of 90Ā° about the z-axis. He then provesāusing geometry and logic onlyāthat the probabilities and, therefore, the magnitudes of the amplitudes (denoted byĀ C+ and Cā and Cā+ and Cāā in the S and T representation respectively) must be the same, but the amplitudes must have different phases, notingāin his typical style, mixing academic and colloquial languageāthat āthere must be some way for a particle to tell that it has turned a corner in (b).ā

The various interpretations of what actually happens here may shed some light on the heated discussions on the reality of the wavefunctionāand of quantum states. In fact, I should note that Feynman’s argument revolves around quantum states. To be precise, the analysis is focused on two-state systems only, and the wavefunctionāwhich captures a continuum of possible states, so to speakāis introduced only later. However, we may look at the amplitude for a particle to be in theĀ up– or down-state as a wavefunction and, therefore (but do note that’s my humble opinion once more), the analysis is actuallyĀ notĀ all that different.

We know, from theory and experiment, that the amplitudes are different. For example, for the given difference in the relative orientation of the two apparatuses (90Ā°), we know that the amplitudes are given by Cā+ = eiāĻ/2āC+ = e iāĻ/4āC+ and Cāā = eāiāĻ/2āC+ = eā iāĻ/4āCā respectively (the amplitude to go from the down to the up state, or vice versa, is zero). Hence, yes, weānotĀ the particle, Mr. Feynman!āknowĀ that, in (b), the electron has, effectively, turned a corner.

The more subtle question here is the following: is the reality of the particle in the two setups the same? Feynman, of course, stays away from such philosophical question. He just notes that, while ā(a) and (b) are differentā, āthe probabilities are the sameā. He refrains from making any statement on the particle itself: is or is it not the same? The common sense answer is obvious: of course, it is! The particle is the same, right? In (b), it just took a turnāso it is just going in some other direction. Thatās all.

However, common sense is seldom a good guide when thinking about quantum-mechanical realities. Also, from a more philosophical point of view, one may argue that the reality of the particle is not the same: something mightāor mustāhave happened to the electron because, when everything is said and done, the particle did take a turn in (b). It did not in (a). [Note that the difference between āmightā and āmustā in the previous phrase may well sum up the difference between a deterministic and a non-deterministic world view but… Well… This discussion is going to be way too philosophical already, so let’s refrain from inserting new language here.]

Let us think this through. The (a) and (b) set-up are, obviously, different but…Ā Wait a minute…Ā Nothing is obvious in quantum mechanics, right? How can weĀ experimentally confirmĀ thatĀ they are different?

Huh?Ā I must be joking, right? You canĀ seeĀ they are different, right? No.Ā I am not joking. In physics, two things are different if we get differentĀ measurementĀ results. [That’s a bit of a simplified view of the ontological point of view of mainstream physicists, but you will have to admit I am not far off.] So… Well… We can’t see those amplitudes and so… Well… If we measure the same thingāsame probabilities, remember?āwhy are they different? Think of this: if we look at the two beam splitters as one singleĀ tube (anĀ ST tube, we might say), then all we did in (b) was bend the tube. Pursuing the logic that says our particle is still the sameĀ even when it takes a turn, we could say the tube is still the same, despite us having wrenched it over a 90Ā° corner.

Now, I am sure you think I’ve just gone nuts, but just tryĀ to stick with me a little bit longer. Feynman actually acknowledges the same: we need to experimentallyĀ proveĀ (a) and (b) are different. He does so by getting aĀ thirdĀ apparatus in (U), as shown below, whose relative orientation to T is the same in both (a) and (b), so there is no difference there.

Now, the axis ofĀ UĀ is not theĀ z-axis: it is theĀ x-axis in (a), and theĀ y-axis in (b). So what? Well… I will quote Feynman hereānot (only) because his words are more important than mine but also because every word matters here:

“The two apparatuses in (a) and (b) are, in fact, different, as we can see in the following way. Suppose that we put an apparatus in front ofĀ SĀ which produces a pure +xĀ state. Such particles would be split into +z andĀ āz intoĀ beams inĀ S,Ā but the two beams would be recombined to give aĀ +xĀ state again at P1āthe exit ofĀ S.Ā The same thing happens again inĀ T.Ā If we followĀ TĀ by a third apparatusĀ U,Ā whose axis is in the +xĀ direction and, as shown in (a), all the particles would go into the +Ā beam ofĀ U.Ā Now imagine what happens ifĀ TĀ and UĀ are swung aroundĀ togetherĀ by 90Ā°Ā to the positions shown in (b).Ā Again, theĀ TĀ apparatus puts out just what it takes in, so the particles that enterĀ UĀ are in a +xĀ stateĀ with respect toĀ S,Ā which is different. By symmetry, we would now expect only one-half of the particles to get through.”

I should note that (b) shows theĀ UĀ apparatus wide open so… Well… I must assume that’s a mistake (and should alert the current editors of the LecturesĀ to it): Feynman’s narrative tells us we should also imagine it with theĀ minus channel shut. InĀ thatĀ case, it should, effectively, filter approximately half of the particles out, while they all get through in (a). So that’s aĀ measurementĀ result which shows the direction, as weĀ seeĀ it, makes a difference.

Now, Feynman would be very angry with meābecause, as mentioned, he hates philosophersābut I’d say: this experiment proves that a direction is something real. Of course, the next philosophical question then is: whatĀ isĀ a direction? I could answer this by pointing to the experiment above: a direction is something that alters the probabilities between the STU tube as set up in (a) versus the STU tube in (b). In factābut, I admit, that would be pretty ridiculousāwe could use the varying probabilities as we wrench this tube over varying angles toĀ define an angle! But… Well… While that’s a perfectly logical argument, I agree it doesn’t sound very sensical.

OK. Next step. What follows may cause brain damage. š Please abandon all pre-conceived notions and definitions for a while and think through the following logic.

You know this stuff is about transformations of amplitudes (or wavefunctions), right? [And you also want to hear about those special 720Ā° symmetry, right? No worries. We’ll get there.] So the questions all revolve around this: what happens to amplitudes (or the wavefunction) when we go from one reference frameāorĀ representation, as it’s referred to in quantum mechanicsāto another?

Well… I should immediately correct myself here: a reference frame and a representation are two different things. They areĀ relatedĀ but… Well… Different… Quite different. Not same-same but different. š I’ll explain why later. Let’s go for it.

Before talking representations, let us first think about what we reallyĀ mean by changing the reference frame. To change it, we first need to answer the question: what is our reference frame? It is a mathematical notion, of course, but then it is also more than that: it is ourĀ reference frame. We use it to make measurements. That’s obvious, you’ll say, but let me make a more formal statement here:

The reference frame is given by (1) the geometry (or theĀ shape, if that sounds easier to you) of the measurement apparatusĀ (so that’s the experimental set-up) here) and (2) our perspective of it.

If we would want to sound academic, we might refer to Kant and other philosophers here, who told usā230 years agoāthat the mathematical idea of a three-dimensional reference frame is grounded in our intuitive notions of up and down, and left and right. [If you doubt this, think about the necessity of the various right-hand rules and conventions that we cannot do without in math, and in physics.] But so we do not want to sound academic. Let us be practical. Just think about the following.Ā The apparatus gives us two directions:

(1) TheĀ upĀ direction, whichĀ weĀ associate with theĀ positive direction of theĀ z-axis, and

(2) the direction of travel of our particle, whichĀ we associateĀ with the positive direction of theĀ y-axis.

Now, if we have two axes, then the third axis (theĀ x-axis) will be given by the right-hand rule, right? So we may say the apparatus gives us the reference frame. Full stop.Ā So… Well… Everything is relative? Is this reference frame relative? Are directions relative? That’s what you’ve been told, but think about this:Ā relativeĀ to what?Ā Here is where the object meets the subject. What’s relative? What’s absolute?Ā Frankly, I’ve started to think that, in this particular situation, we should, perhaps, not use these two terms. I am notĀ saying thatĀ our observation of what physically happens here gives these two directions any absolute character but… Well… You will have to admit they are more than just some mathematical construct: when everything is said and done, we will have to admit that these two directions are real. because… Well… They’re part of theĀ realityĀ that we are observing, right? And the third one… Well… That’s given by our perspectiveāby our right-hand rule, which is… Well… OurĀ right-hand rule.

Of course, now you’ll say: if you think that ārelativeā and āabsoluteā are ambiguous terms and that we, therefore, may want to avoid them a bit more, then ārealā and its opposite (unreal?) are ambiguous terms too, right? Wellā¦ Maybe. What language would youĀ suggest? š Just stick to the story for a while. I am not done yet. So… Yes… WhatĀ isĀ theirĀ reality?Ā Let’s think about that in the next section.

### Perspectives, reference frames and symmetries

You’ve done some mental exercises already as you’ve been working your way through the previous section, but you’ll need to do plenty more. In fact, they may become physical exercise too: when I first thought about these things (symmetries and, more importantly, asymmetries in space), I found myself walking around the table with some asymmetrical everyday objects and papers with arrows and clocks and other stuff on itāeffectively analyzing what right-hand screw, thumb or grip rules actuallyĀ mean. š

So… Well… I want you to distinguishājust for a whileābetween the notion of a reference frame (think of the xyz reference frame that comes with the apparatus) and yourĀ perspective on it. What’s our perspective on it? Well… You may be looking from the top, or from the side and, if from the side, from the left-hand side or the right-hand sideāwhich, if you think about it, you can only defineĀ in terms of the various positive and negative directions of the various axes. šĀ If you think this is getting ridiculous… Well… Don’t. Feynman himselfĀ doesn’t think this is ridiculous, because he starts his own “long and abstract side tour” on transformations with a very simple explanation of how the top and side view of the apparatus are related to theĀ axesĀ (i.e. the reference frame) that comes with it. You don’t believe me? This is theĀ very first illustration of hisĀ LectureĀ on this:

He uses it to explain the apparatus (which we don’t do here because you’re supposed to already know how these (modified or improved) Stern-Gerlach apparatuses work). So let’s continue this story. Suppose that we are looking in the positiveĀ y-directionāso thatās the direction in which our particle is movingāthen we might imagine how it would look like whenĀ weĀ would make a 180Ā°Ā turn and look at the situation from the other side, so to speak. We do not change the reference frame (i.e. the orientation) of the apparatus here: we just change our perspective on it. Instead of seeing particles going away from us, into the apparatus, we now see particles comingĀ towardsĀ us, out of the apparatus.

What happensābut that’s not scientific language, of courseāis that left becomes right, and right becomes left. Top still is top, and bottom is bottom. We are looking now in theĀ negativeĀ y-direction, and the positive direction of the x-axisāwhich pointed right when we were looking in the positiveĀ y-directionānow points left. I see you nodding your head nowābecause you’ve heard about parity inversions, mirror symmetries and what have youāand I hear you say: “That’s the mirror world, right?”

No. It is not. I wrote about this in another post: the world in the mirror is theĀ world in the mirror. We don’t get a mirror image of an object by going around it and looking at its back side. I can’t dwell too much on this (just check that post, and another one who talks about the same), but so don’t try to connect it to the discussions on symmetry-breaking and what have you. Just stick toĀ this story, which is about transformations of amplitudes (or wavefunctions). [If you really want to knowābut I know this sounds counterintuitiveāthe mirror world doesn’t really switch left for right. Your reflection doesn’t do a 180 degree turn: it is just reversed front to back, with no rotation at all. It’s only your brain which mentallyĀ adds (or subtracts) the 180 degree turn that you assume must have happened from the observed front to back reversal. So the left to right reversal is onlyĀ apparent. It’s a common misconception, and… Well… I’ll let you figure this out yourself. I need to move on.]Ā Just note the following:

1. TheĀ xyzĀ reference frame remains a valid right-handed reference frame. Of course it does: it comes with our beam splitter, and we can’t change its reality, right? We’re just looking at it from another angle. OurĀ perspectiveĀ on it has changed.
2. However, if we think of the real and imaginary part of the wavefunction describing the electrons that are going through our apparatus as perpendicular oscillations (as shown below)āa cosine and sine function respectivelyāthen our change in perspectiveĀ might, effectively, mess up our convention for measuring angles.

I am not saying itĀ does. Not now, at least. I am just saying it might. It depends on the plane of the oscillation, as I’ll explain in a few moments. Think of this: we measure angles counterclockwise, right? As shown below… But… Well… If the thing below would be some funny clock going backwardsāyou’ve surely seen them in a bar or so, right?āthen… Well… If they’d be transparent, and you’d go around them, you’d see them as going… Yes… Clockwise. š [This should remind you of a discussion on real versus pseudo-vectors, or polar versus axial vectors, but… Well… We don’t want to complicate the story here.]

Now, ifĀ we wouldĀ assume this clock represents something realāand, of course, I am thinking of theĀ elementary wavefunctionĀ eiĪøĀ =Ā cosĪø +Ā iĀ·sinĪø nowāthen… Well… Then it will look different when we go around it. When going around our backwards clock above and looking at it from… Well… The back, we’d describe it, naively, as… Well…Ā Think! What’s your answer? Give me the formula!Ā š

[…]

We’d see it asĀ eāiĪøĀ =Ā cos(āĪø) +Ā iĀ·sin(āĪø) =Ā cosĪø āĀ iĀ·sinĪø, right? The hand of our clock now goes clockwise, so that’s theĀ oppositeĀ direction of our convention for measuring angles. Hence, instead ofĀ eiĪø, we writeĀ eāiĪø, right? So that’s the complex conjugate. So we’ve got a differentĀ imageĀ of the same thing here. Not good. Not good at all.

You’ll say: so what? We can fix this thing easily, right?Ā YouĀ don’t need the convention for measuring angles or for the imaginary unit (i) here.Ā This particle is moving, right? So if you’d want to look at the elementary wavefunction as some sort of circularly polarized beam (which, I admit, is very much what I would like to do, but its polarization is rather particular as I’ll explain in a minute), then you just need to define left- and right-handed angles as per the standard right-hand screw rule (illustrated below).Ā To hell with the counterclockwise convention for measuring angles!

You are right. WeĀ couldĀ use the right-hand rule more consistently. We could, in fact, use it as anĀ alternativeĀ convention for measuring angles: we could, effectively, measure them clockwise or counterclockwise depending on the direction of our particle.Ā But… Well… The fact is:Ā we don’t. We do not use that alternative convention when we talk about the wavefunction. Physicists do use theĀ counterclockwiseĀ convention all of the time and just jot down these complex exponential functions and don’t realize that,Ā if they are to represent something real, ourĀ perspectiveĀ on the reference frame matters. To put it differently, theĀ directionĀ in which we are looking at things matters! Hence, the direction is not…Ā Well… I am tempted to say… NotĀ relative at all but then… Well… We wanted to avoid that term, right? š

[…]

I guess that, by now, your brain may suffered from various short-circuits. If not, stick with me a while longer. Let us analyze how our wavefunction model might be impacted by this symmetryāorĀ asymmetry, I should say.

### The flywheel model of an electron

In our previous posts, we offered a model that interprets the real and the imaginary part of the wavefunction as oscillations which each carry half of the total energy of the particle. These oscillations are perpendicular to each other, and the interplay between both is how energy propagates through spacetime. Let us recap the fundamental premises:

1. The dimension of the matter-wave field vector is forceĀ per unit mass (N/kg), as opposed to the force per unit charge (N/C) dimension of the electric field vector. This dimension is an acceleration (m/s2), which is the dimension of the gravitational field.
2. We assume this gravitational disturbance causes our electron (or a charged massĀ in general) to move about some center, combining linear and circular motion. This interpretation reconciles the wave-particle duality: fields interfere but if, at the same time, they do drive a pointlike particle, then we understand why, as Feynman puts it, āwhen you do find the electron some place, the entire charge is there.ā Of course, we cannot prove anything here, but our elegant yet simple derivation of the Compton radius of an electron is… Well… Just nice. š
3. Finally, and most importantly in the context of this discussion, we noted that, in light of the direction of the magnetic moment of an electron in an inhomogeneous magnetic field, the plane which circumscribes the circulatory motion of the electron should also compriseĀ the direction of its linear motion. Hence, unlike an electromagnetic wave, theĀ planeĀ of the two-dimensional oscillation (so that’s the polarization plane, really) cannotĀ be perpendicular to the direction of motion of our electron.

Let’s say some more about the latter point here. The illustrations below (one from Feynman, and the other is just open-source) show what we’re thinking of.Ā The direction of the angular momentum (and the magnetic moment) of an electronāor, to be precise, its component as measured in the direction of the (inhomogeneous) magnetic field through which our electron is travelingācannotĀ be parallel to the direction of motion. On the contrary, it must be perpendicularĀ to the direction of motion. In other words, if we imagine our electron as spinning around some center (see the illustration on the left-hand side), then the disk it circumscribes (i.e. theĀ planeĀ of the polarization)Ā has toĀ compriseĀ the direction of motion.

Of course, we need to add another detail here. As my readers will know, we do not really have a precise direction of angular momentum in quantum physics. While there is no fully satisfactory explanation of this, the classical explanationācombined with the quantization hypothesisāgoes a long way in explaining this: an object with an angular momentumĀ JĀ and a magnetic momentĀ Ī¼Ā that is not exactly parallel to some magnetic fieldĀ B, willĀ notĀ line up: it willĀ precessāand, as mentioned, the quantization of angular momentum may well explain the rest.Ā [Well… Maybe… We haveĀ detailed our attempts in this regard in various posts on this (just search for spinĀ orĀ angular momentumĀ on this blog, and you’ll get a dozen posts or so), but these attempts are, admittedly, not fully satisfactory. Having said that, they do go a long way in relating angles to spin numbers.]

The thing is: we do assume our electron is spinning around. If we look from theĀ up-direction only, then it will be spinningĀ clockwise if its angular momentum is down (so itsĀ magnetic moment isĀ up). Conversely, it will be spinningĀ counterclockwise if its angular momentum isĀ up. Let us take theĀ up-state. So we have a top view of the apparatus, and we see something like this:I know you are laughing aloud now but think of your amusement as a nice reward for having stuck to the story so far. Thank you. š And, yes, do check it yourself by doing some drawings on your table or so, and then look at them from various directions as you walk around the table asāI am not ashamed to admit thisāI did when thinking about this. So what do we get when we change the perspective? Let us walk around it, counterclockwise, let’s say, so we’re measuring our angle of rotation as someĀ positiveĀ angle.Ā Walking around itāin whatever direction, clockwise or counterclockwiseādoesn’t change the counterclockwise direction of our… Well… That weird object that mightājust mightārepresent an electron that has its spin up and that is traveling in the positive y-direction.

When we look in the direction of propagation (so that’s from left to right as you’re looking at this page), and we abstract away from its linear motion, then we could, vaguely, describe this by some wrenchedĀ eiĪøĀ =Ā cosĪø +Ā iĀ·sinĪø function, right? The x- andĀ y-axesĀ of the apparatus may be used to measure the cosine and sine components respectively.

Let us keep looking from the top but walk around it, rotating ourselves over a 180Ā° angle so we’re looking in theĀ negativeĀ y-direction now. As I explained in one of those posts on symmetries, our mind will want to switch to a new reference frame: we’ll keep theĀ z-axis (up is up, and down is down), but we’ll want the positive direction of the x-axis to… Well… Point right. And we’ll want theĀ y-axis to point away, rather than towards us. In short, we have a transformation of the reference frame here:Ā z’ =Ā z,Ā y’ = āĀ y, andĀ x’ =Ā āĀ x. Mind you, this is still a regular right-handed reference frame. [That’s the difference with aĀ mirrorĀ image: aĀ mirroredĀ right-hand reference frame is no longer right-handed.]Ā So, in our new reference frame, that we choose to coincide with ourĀ perspective,Ā we will now describe the same thing as someĀ ācosĪø āĀ iĀ·sinĪø =Ā āeiĪøĀ function. Of course,Ā ācosĪø =Ā cos(Īø +Ā Ļ) andĀ āsinĪø =Ā sin(Īø +Ā Ļ) so we can write this as:

ācosĪø āĀ iĀ·sinĪø =Ā cos(Īø +Ā Ļ) +Ā iĀ·sinĪø =Ā eiĀ·(Īø+Ļ)Ā =Ā eiĻĀ·eiĪøĀ = āeiĪø.

Sweet ! But… Well… First note this isĀ notĀ the complex conjugate:Ā eāiĪøĀ =Ā cosĪø āĀ iĀ·sinĪøĀ ā Ā ācosĪø āĀ iĀ·sinĪø =Ā āeiĪø. Why is that? Aren’t we looking at the same clock, but from the back? No. The plane of polarization is different. Our clock is more like those in Dali’s painting: it’s flat. š And, yes, let me lighten up the discussion with that painting here. š We need to haveĀ someĀ fun while torturing our brain, right?

So, because we assume the plane of polarization is different, we get anĀ āeiĪøĀ function instead of aĀ eāiĪøĀ function.

Let us now think about the eiĀ·(Īø+Ļ)Ā function. It’s the same asĀ āeiĪøĀ but… Well… We walked around theĀ z-axis taking a full 180Ā° turn, right? So that’s Ļ in radians. So that’s the phase shiftĀ here. Hey!Ā Try the following now. Go back and walk around the apparatus once more, but letĀ the reference frame rotate with us, as shown below. So we start left and look in the direction of propagation, and then we start moving about theĀ z-axis (which points out of this page, toward you, as you are looking at this), let’s say by some small angleĀ Ī±. So we rotate the reference frame about theĀ z-axis byĀ Ī± and… Well… Of course, ourĀ eiĀ·ĪøĀ now becomes anĀ ourĀ eiĀ·(Īø+Ī±)Ā function, right? We’ve just derived the transformation coefficient for a rotation about theĀ z-axis, didn’t we? It’s equal toĀ eiĀ·Ī±, right? We get the transformed wavefunction in the new reference frame by multiplying the old one byĀ eiĀ·Ī±, right? It’s equal toĀ eiĀ·Ī±Ā·eiĀ·ĪøĀ =Ā eiĀ·(Īø+Ī±), right?

Well…

[…]

No. The answer is: no. TheĀ transformation coefficient is notĀ eiĀ·Ī±Ā butĀ eiĀ·Ī±/2. So we get an additional 1/2 factor in theĀ phase shift.

Huh?Ā Yes.Ā That’s what it is: when we change the representation, by rotating our apparatus over some angle Ī± about the z-axis, then we will, effectively, get a new wavefunction, which will differ from the old one by a phase shift that is equal to onlyĀ half ofĀ the rotation angle only.

Huh?Ā Yes. It’s even weirder than that. For a spin downĀ electron, the transformation coefficient is eāiĀ·Ī±/2, so we get an additional minus sign in the argument.

Huh?Ā Yes.

I know you are terribly disappointed, but that’s how it is. That’s what hampers an easy geometric interpretation of the wavefunction. Paraphrasing Feynman, I’d say that, somehow, our electron not only knows whether or not it has taken a turn, but it also knows whether or not it is moving away from us or, conversely, towards us.

[…]

But…Ā Hey! Wait a minute! That’s it, right?Ā

What? Well… That’s it! The electron doesn’t know whether it’s moving away or towards us. That’s nonsense. But… Well… It’s like this:

OurĀ eiĀ·Ī±Ā coefficient describes a rotation of the reference frame. In contrast, theĀ eiĀ·Ī±/2Ā andĀ eāiĀ·Ī±/2Ā coefficients describe what happens when we rotate the T apparatus! Now thatĀ is a very different proposition.Ā

Right! You got it! RepresentationsĀ and reference frames are different things.Ā QuiteĀ different, I’d say: representations areĀ real, reference frames aren’tābut then you don’t like philosophical language, do you? šĀ But think of it. When we just go about theĀ z-axis, a full 180Ā°, but we don’t touch thatĀ T-apparatus, we don’t changeĀ reality. When we were looking at the electron while standing left to the apparatus, we watched the electrons going in and moving away from us, and when we go about theĀ z-axis, a full 180Ā°, looking at it from the right-hand side, we see the electrons coming out, moving towards us. But it’s still the same reality. We simply change the reference frameāfrom xyz to x’y’z’ to be precise: we doĀ not changeĀ the representation.

In contrast, when we rotate theĀ TĀ apparatus over a full 180Ā°, our electron now goes in the opposite direction. And whether that’s away or towards us, that doesn’t matter: it was going in one direction while traveling throughĀ S, and now it goes in the opposite directionārelative to the direction it was going in S, that is.

So what happens,Ā really, when weĀ change the representation, rather than the reference frame? Well… Let’s think about that. š

### Quantum-mechanical weirdness?

The transformation matrix for the amplitude of a system to be in anĀ upĀ orĀ downĀ state (and, hence, presumably, for a wavefunction) for a rotation about theĀ z-axis is the following one:

Feynman derives this matrix in a rather remarkable intellectualĀ tour de forceĀ in the 6th of hisĀ Lectures on Quantum Mechanics. So that’s pretty early on. He’s actually worried about that himself, apparently, and warns his students that “This chapter is a rather long and abstract side tour, and it does not introduce any idea which we will not also come to by a different route in later chapters. You can, therefore, skip over it, and come back later if you are interested.”

Well… That’s howĀ IĀ approached it. I skipped it, and didn’t worry about those transformations for quite a while. But… Well… You can’t avoid them. In some weird way, they are at the heart of the weirdness of quantum mechanics itself. Let us re-visit his argument. Feynman immediately gets that the whole transformation issue here is just a matter of finding an easy formula for that phase shift. Why? He doesn’t tell us. Lesser mortals like us must just assume that’s how the instinct of a genius works, right? š So… Well… Because heĀ knowsāfrom experimentāthat the coefficient isĀ eiĀ·Ī±/2Ā instead of eiĀ·Ī±, he just says the phase shiftāwhich he denotes by Ī»āmust be someĀ proportionalĀ to the angle of rotationāwhich he denotes byĀ Ļ rather than Ī± (so as to avoid confusion with the EulerĀ angleĀ Ī±). So he writes:

Ī» =Ā mĀ·Ļ

Initially, he also tries the obvious thing: m should be one, right? SoĀ Ī» = Ļ, right? Well… No. It can’t be. Feynman shows why that can’t be the case by adding a third apparatus once again, as shown below.

Let me quote him here, as I can’t explain it any better:

“SupposeĀ TĀ is rotated byĀ 360Ā°; then, clearly, it is right back at zero degrees, and we should haveĀ Cā+ = C+Ā andĀ Cāā =Ā CāĀ or,Ā what is the same thing,Ā eiĀ·mĀ·2ĻĀ = 1. We get m =Ā 1. [But no!]Ā This argument is wrong!Ā To see that it is, consider thatĀ TĀ is rotated byĀ 180Ā°. If mĀ were equal to 1, we would have Cā+ =Ā eiĀ·ĻC+Ā = āC+Ā and Cāā =Ā eāiĀ·ĻCāĀ =Ā āCā. [Feynman works with statesĀ here, instead of the wavefunction of the particle as a whole. I’ll come back to this.] However, this is just theĀ originalĀ state all over again.Ā BothĀ amplitudes are just multiplied byĀ ā1Ā which gives back the original physical system. (It is again a case of a common phase change.) This means that if the angle betweenĀ TĀ andĀ SĀ is increased to 180Ā°, the system would be indistinguishable from the zero-degree situation, and the particles would again go through the (+)Ā state of theĀ UĀ apparatus. AtĀ 180Ā°, though, the (+)Ā state of theĀ UĀ apparatus is theĀ (āx)Ā state of the originalĀ SĀ apparatus. So a (+x)Ā state would become aĀ (āx)Ā state. But we have done nothing toĀ changeĀ the original state; the answer is wrong. We cannot haveĀ m = 1.Ā We must have the situation that a rotation byĀ 360Ā°, andĀ no smaller angleĀ reproduces the same physical state. This will happen ifĀ m = 1/2.”

The result, of course, is this weird 720Ā° symmetry. While we get the same physics after a 360Ā° rotation of the T apparatus, we doĀ notĀ get the same amplitudes. We get the opposite (complex) number:Ā Cā+ =Ā eiĀ·2Ļ/2C+Ā = āC+Ā and Cāā =Ā eāiĀ·2Ļ/2CāĀ =Ā āCā. That’s OK, because… Well… It’s aĀ commonĀ phase shift, so it’s just like changing the origin of time. Nothing more. Nothing less. Same physics. Same reality. But… Well…Ā Cā+ ā Ā āC+Ā andĀ Cāā ā Ā āCā, right? We only get our original amplitudes back if we rotate theĀ T apparatus two times, so that’s by a full 720 degreesāas opposed to the 360Ā° we’d expect.

Now, space is isotropic, right? So this 720Ā° business doesn’t make sense, right?

Well… It does and it doesn’t. We shouldn’t dramatize the situation. What’s the actual difference between a complex number and its opposite? It’s like x orĀ āx, or t and āt.Ā I’ve said this a couple of times already again, and I’ll keep saying it many times more:Ā NatureĀ surely can’t be bothered by how we measure stuff, right? In the positive or the negative directionāthat’s just our choice, right?Ā OurĀ convention. So… Well… It’s just like thatĀ āeiĪøĀ function we got when looking at theĀ same experimental set-up from the other side: ourĀ eiĪøĀ and āeiĪøĀ functions didĀ notĀ describe a different reality. We just changed our perspective. TheĀ reference frame. As such, the reference frame isn’tĀ real. The experimental set-up is. AndāI know I will anger mainstream physicists with thisātheĀ representationĀ is. Yes. Let me say it loud and clear here:

A different representation describes a different reality.

In contrast, a different perspectiveāor a different reference frameādoes not.

### Conventions

While you might have had a lot of trouble going through all of the weird stuff above, the point is: it isĀ notĀ all that weird. WeĀ canĀ understand quantum mechanics. And in a fairly intuitive way, really. It’s just that… Well… I think some of the conventions in physics hamper such understanding. Well… Let me be precise: one convention in particular, really. It’s that convention for measuring angles. Indeed, Mr. Leonhard Euler, back in the 18th century, might well be “the master of us all” (as Laplace is supposed to have said) but… Well… He couldn’t foresee how his omnipresent formulaāeiĪøĀ =Ā cosĪø +Ā iĀ·sinĪøāwould, one day, be used to representĀ something real: an electron, or any elementary particle, really. If he wouldĀ have known, I am sure he would have noted what I am noting here:Ā NatureĀ can’t be bothered by our conventions. Hence, ifĀ eiĪøĀ represents something real, thenĀ eāiĪøĀ must also represent something real. [Coz I admire this genius so much, I can’t resist the temptation. Here’s his portrait. He looks kinda funny here, doesn’t he? :-)]

Frankly, he would probably have understood quantum-mechanical theory as easily and instinctively as Dirac, I think, and I am pretty sure he would have notedāand, if he would have known about circularly polarized waves, probably agreed toāthatĀ alternative convention for measuring angles: we could, effectively, measure angles clockwise or counterclockwise depending on the direction of our particleāas opposed to Euler’s ‘one-size-fits-all’ counterclockwise convention. But so we didĀ notĀ adopt that alternative convention because… Well… We want to keep honoring Euler, I guess. š

So… Well… If we’re going to keep honoring Euler by sticking to that ‘one-size-fits-all’ counterclockwise convention, then I doĀ believe thatĀ eiĪøĀ and eāiĪøĀ represent twoĀ differentĀ realities: spin up versus spin down.

Yes. In our geometric interpretation of the wavefunction, these are, effectively, two different spin directions. And… Well… These are real directions: we seeĀ something different when they go through a Stern-Gerlach apparatus. So it’s not just some convention toĀ countĀ things like 0, 1, 2, etcetera versus 0,Ā ā1,Ā ā2 etcetera. It’s the same story again: different but relatedĀ mathematicalĀ notions are (often) related to different but relatedĀ physicalĀ possibilities. So… Well… I think that’s what we’ve got here.Ā Think of it. Mainstream quantum math treats all wavefunctions as right-handed but… Well…Ā A particle with up spin is a different particle than one withĀ downĀ spin, right? And, again,Ā NatureĀ surely cannotĀ be bothered about our convention of measuring phase angles clockwise or counterclockwise, right? So… Well… Kinda obvious, right? š

Let me spell out my conclusions here:

1. The angular momentum can be positive or, alternatively, negative: J = +Ä§/2 orĀ āÄ§/2. [Let me note that this is not obvious. Or less obvious than it seems, at first. In classical theory, you would expect an electron, or an atomic magnet, to line up with the field. Well… The Stern-Gerlach experiment shows they don’t: they keep their original orientation. Well… If the field is weak enough.]

2. Therefore, we would probably like to think that an actual particleāthink of an electron, or whatever other particle you’d think ofācomes in twoĀ variants:Ā right-handed and left-handed. They will, therefore,Ā either consist of (elementary) right-handed waves or,Ā else, (elementary) left-handed waves. An elementary right-handed wave would be written as: Ļ(Īøi)Ā = eiĪøiĀ = aiĀ·(cosĪøi + iĀ·sinĪøi). In contrast,Ā an elementary left-handed wave would be written as: Ļ(Īøi)Ā =Ā eāiĪøiĀ = aiĀ·(cosĪøi ā iĀ·sinĪøi).Ā So that’s the complex conjugate.

So… Well… Yes, I think complex conjugates are not just someĀ mathematicalĀ notion: I believe they represent something real. It’s the usual thing:Ā NatureĀ has shown us that (most) mathematical possibilities correspond to realĀ physical situations so… Well… Here you go. It is reallyĀ just like the left- or right-handed circular polarization of an electromagnetic wave: we can have both for the matter-wave too! [As for the differencesādifferent polarization plane and dimensions and what have youāI’ve already summed those up, so I won’t repeat myself here.]Ā The point is: ifĀ we have two differentĀ physicalĀ situations, we’ll want to have two different functions to describe it. Think of it like this: why would we haveĀ twoāyes, I admit, two relatedāamplitudes to describe the upĀ or downĀ state of the same system, but only one wavefunction for it?Ā You tell me.

[…]

Authors like me are looked down upon by the so-called professional class of physicists. The few who bothered to react to my attempts to make sense of Einstein’s basic intuition in regard to the nature of the wavefunction all said pretty much the same thing: “Whatever your geometric (orĀ physical) interpretation of the wavefunction might be, it won’t be compatible with theĀ isotropyĀ of space. You cannot imagineĀ an object with a 720Ā° symmetry. That’sĀ geometrically impossible.”

Well… Almost three years ago, I wrote the following on this blog: “As strange as it sounds, aĀ spin-1/2 particle needsĀ twoĀ full rotations (2Ć360Ā°=720Ā°) until it is again in the same state. Now, in regard to that particularity, youāll often read something like: āThere isĀ nothingĀ in our macroscopic world which has a symmetry like that.ā Or, worse, āCommon sense tells us that something like that cannot exist, that it simply is impossible.ā [I wonāt quote the site from which I took this quotes, because it is, in fact, the site of a very respectable Ā research center!]Ā Bollocks!Ā TheĀ Wikipedia article on spinĀ has this wonderful animation: look at how the spirals flip between clockwise and counterclockwise orientations, and note that itās only after spinning a full 720 degrees that this āpointā returns to its original configuration after spinning a full 720 degrees.

So… Well… I am still pursuing my original dream which is… Well… Let me re-phrase what I wrote back in January 2015:

Yes, weĀ canĀ actually imagine spin-1/2 particles, and we actually do not need all that much imagination!

In fact, I am tempted to think that I’ve found a pretty good representation or… Well… A pretty goodĀ image, I should say, because… Well… A representation is something real, remember? š

Post scriptum (10 December 2017):Ā Our flywheel model of an electron makes sense, but also leaves many unanswered questions. The most obvious one question, perhaps, is: why theĀ upĀ andĀ downĀ state only?

I am not so worried about that question, even if I can’t answer it right away because… Well… Our apparatusāthe way weĀ measureĀ realityāis set up to measure the angular momentum (or the magnetic moment, to be precise) in one direction only. If our electron isĀ capturedĀ by someĀ harmonicĀ (or non-harmonic?) oscillation in multiple dimensions, then it should not be all that difficult to show its magnetic moment is going to align, somehow, in the same or, alternatively, the opposite direction of the magnetic field it is forced to travel through.

Of course, the analysis for the spinĀ upĀ situation (magnetic moment down) is quite peculiar: if our electron is aĀ mini-magnet, why would itĀ notĀ line up with the magnetic field? We understand the precession of a spinning top in a gravitational field, but…Ā Hey… It’s actually not that different. Try to imagine some spinning top on the ceiling. š I am sure we can work out the math. š The electron must be some gyroscope, really: it won’t change direction. In other words, its magnetic moment won’t line up. It will precess, and it can do so in two directions, depending on its state. š […] At least, that’s why my instinct tells me. I admit I need to work out the math to convince you. š

The second question is more important. If we just rotate the reference frame over 360Ā°, we see the same thing: some rotating object which we, vaguely, describe by someĀ e+iĀ·ĪøĀ functionāto be precise, I should say: by some Fourier sum of such functionsāor, if the rotation is in the other direction, by someĀ eāiĀ·ĪøĀ function (again, you should read: aĀ FourierĀ sum of such functions). Now, the weird thing, as I tried to explain above is the following: if we rotate the object itself, over the sameĀ 360Ā°, we get aĀ differentĀ object: ourĀ eiĀ·ĪøĀ andĀ eāiĀ·ĪøĀ function (again: think of aĀ FourierĀ sum, so that’s a waveĀ packet, really) becomes aĀ āeĀ±iĀ·ĪøĀ thing. We get aĀ minusĀ sign in front of it.Ā So what happened here? What’s the difference, really?

Well… I don’t know. It’s very deep. If I do nothing, and you keep watching me while turning around me, for a fullĀ 360Ā°, then you’ll end up where you were when you started and, importantly, you’ll see the same thing.Ā ExactlyĀ the same thing: if I was anĀ e+iĀ·ĪøĀ wave packet, I am still anĀ anĀ e+iĀ·ĪøĀ wave packet now. OrĀ if I was an eāiĀ·ĪøĀ wave packet, then I am still anĀ an eāiĀ·ĪøĀ wave packet now. Easy. Logical. Obvious, right?

But so now we try something different:Ā IĀ turn around, over a fullĀ 360Ā° turn, and youĀ stay where you are. When I am back where I wasālooking at you again, so to speakāthen… Well… I am not quite the same any more. Or… Well… Perhaps I am but youĀ seeĀ me differently. If I wasĀ e+iĀ·ĪøĀ wave packet, then I’ve become aĀ āe+iĀ·ĪøĀ wave packet now. Not hugely different but… Well… ThatĀ minusĀ sign matters, right? OrĀ If I wasĀ wave packet built up from elementaryĀ aĀ·eāiĀ·ĪøĀ waves, then I’ve become aĀ āeāiĀ·ĪøĀ wave packet now. What happened?

It makes me think of the twin paradox in special relativity. We know it’s aĀ paradoxāso that’s anĀ apparentĀ contradiction only: we know which twin stayed on Earth and which one traveled because of the gravitational forces on the traveling twin. The one who stays on Earth does not experience any acceleration or deceleration. Is it the same here? I mean… The one who’s turning around must experience someĀ force.

Can we relate this to the twin paradox? Maybe. Note that aĀ minusĀ sign in front of theĀ eāĀ±iĀ·ĪøĀ functions amounts a minus sign in front of both the sine and cosine components. So… Well… The negative of a sine and cosine is the sine and cosine but with a phase shift of 180Ā°: ācosĪø =Ā cos(Īø Ā± Ļ) andĀ āsinĪø =Ā sin(Īø Ā± Ļ). Now, adding or subtracting aĀ commonĀ phase factor to/from the argument of the wavefunction amounts toĀ changingĀ the origin of time. So… Well… I do think the twin paradox and this rather weird business of 360Ā° and 720Ā° symmetries are, effectively, related. š

# The reality of the wavefunction

If you haven’t read any of my previous posts on the geometry of the wavefunction (this link goes to the most recent one of them), then don’t attempt to read this one. It brings too much stuff together to be comprehensible. In fact, I am not even sure if I am going to understand what I write myself. š [OK. Poor joke. Acknowledged.]

Just to recap the essentials, I part ways with mainstream physicists in regard to theĀ interpretationĀ of the wavefunction. For mainstream physicists, the wavefunction is just some mathematical construct. NothingĀ real. Of course, I acknowledge mainstream physicists have very good reasons for that, but… Well… I believe that, if there is interference, or diffraction, thenĀ somethingĀ must be interfering, or something must be diffracting. I won’t dwell on this because… Well… I have done that too many times already. MyĀ hypothesisĀ is that the wavefunction is, in effect, aĀ rotatingĀ field vector, so itās just like the electric field vector of a (circularly polarized) electromagnetic wave (illustrated below).

Of course, it must be different, and it is. First, theĀ (physical) dimension of the field vector of the matter-wave must be different. So what is it? Well… I am tempted to associate the real and imaginary component of the wavefunction with a forceĀ per unit massĀ (as opposed to the force per unit charge dimension of the electric field vector). Of course, the newton/kg dimension reduces to the dimension of acceleration (m/s2), so thatās the dimension of a gravitational field.

Second, I also am tempted to think that this gravitational disturbance causes an electron (or any matter-particle) to move about some center, and I believe it does so at the speed of light. In contrast, electromagnetic waves doĀ notĀ involve any mass: theyāre just an oscillatingĀ field. Nothing more. Nothing less. Why would I believe there must still be some pointlike particle involved? Well…Ā As Feynman puts it: āWhen you do find the electron some place, the entire charge is there.ā (FeynmanāsĀ Lectures, III-21-4) So… Well… That’s why.

The third difference is one that I thought of only recently: theĀ planeĀ of the oscillation cannotĀ be perpendicular to the direction of motion of our electron, because then we canāt explain the direction of its magnetic moment, which is either up or down when traveling through a Stern-Gerlach apparatus. I am more explicit on that in the mentioned post, so you may want to check there. š

I wish I mastered the software to make animations such as the one above (for which I have to credit Wikipedia), but so I don’t. You’ll just have toĀ imagineĀ it. That’s great mental exercise, so… Well… Just try it. š

Let’s now think about rotating reference frames and transformations. If theĀ z-direction is the direction along which we measure the angular momentum (or the magnetic moment), then theĀ up-direction will be theĀ positiveĀ z-direction. We’ll also assume theĀ y-direction is the direction of travel of our elementary particleāand let’s just consider an electron here so we’re moreĀ real. š So we’re in the reference frame that Feynman used to derive the transformation matrices for spin-1/2 particles (or for two-state systems in general). His ‘improved’ Stern-Gerlach apparatusāwhich I’ll refer to as a beam splitterāillustrates this geometry.

So I think the magnetic momentāor the angular momentum, reallyācomes from an oscillatory motion in the x– and y-directions. One is theĀ realĀ component (the cosine function) and the other is the imaginary component (the sine function), as illustrated below.Ā

So the crucial difference with the animations above (which illustrate left- and a right-handed polarization respectively) is that we, somehow, need to imagine the circular motion isĀ notĀ in theĀ xz-plane, but in theĀ yz-plane. Now what happens if we change the reference frame?

Well… That depends on what you mean by changing the reference frame. Suppose we’re looking in the positive y-directionāso that’s the direction in which our particle is movingā, then we might imagine how it would look like whenĀ weĀ would make a 180Ā°Ā turn and look at the situation from the other side, so to speak. Now, I did a post on that earlier this year, which you may want to re-read.Ā When we’re looking at the same thing from the other side (from the back side, so to speak), we will want to use our familiar reference frame. So we will want to keep theĀ z-axis as it is (pointing upwards), and we will also want to define theĀ x– andĀ y-axis using the familiar right-hand rule for defining a coordinate frame. So our newĀ x-axis and our newĀ y-axis will the same as the oldĀ x- andĀ y-axes but with the sign reversed. In short, we’ll have the following mini-transformation: (1)Ā z‘ =Ā z, (2) x’ = āx, and (3) y’ =Ā āy.

So… Well… If we’re effectively looking at somethingĀ realĀ that was moving along theĀ y-axis, then it will now still be moving along the y’-axis, butĀ in theĀ negativeĀ direction. Hence, our elementary wavefunctionĀ eiĪøĀ = cosĪø +Ā iĀ·sinĪø willĀ transformĀ intoĀ ācosĪø āĀ iĀ·sinĪø =Ā ācosĪø āĀ iĀ·sinĪø =Ā cosĪø āĀ iĀ·sinĪø.Ā It’s the same wavefunction. We just… Well… We just changed our reference frame. We didn’t change reality.

Now you’ll cry wolf, of course, because we just went through all that transformational stuff in our last post. To be specific, we presented the following transformation matrix for a rotation along theĀ z-axis:

Now, ifĀ Ļ is equal to 180Ā° (so that’s Ļ in radians), then theseĀ eiĻ/2Ā andĀ eāiĻ/2/ā2Ā factors areĀ equal toĀ eiĻ/2Ā =Ā +iĀ andĀ eāiĻ/2Ā = āiĀ respectively. Hence, ourĀ eiĪøĀ = cosĪø +Ā iĀ·sinĪø becomes…

Hey ! Wait a minute ! We’re talking about twoĀ veryĀ different things here, right? TheĀ eiĪøĀ = cosĪø +Ā iĀ·sinĪø is anĀ elementaryĀ wavefunction which, we presume, describes some real-life particleāwe talked about an electron with its spin in theĀ up-directionāwhile these transformation matrices are to be applied to amplitudes describing… Well… Either anĀ up– or a down-state, right?

Right. But… Well… Is itĀ so different, really? Suppose ourĀ eiĪøĀ = cosĪø +Ā iĀ·sinĪø wavefunction describes anĀ up-electron, then we still have to apply thatĀ eiĻ/2Ā =Ā eiĻ/2Ā =Ā +iĀ factor, right? So we get a new wavefunction that will be equal toĀ eiĻ/2Ā·eiĪøĀ =Ā eiĻ/2Ā·eiĪøĀ =Ā +iĀ·eiĪøĀ =Ā iĀ·cosĪø +Ā i2Ā·sinĪø =Ā sinĪø āĀ iĀ·cosĪø, right? So how can we reconcile that with the cosĪø āĀ iĀ·sinĪø function we thought we’d find?

We can’t. So… Well… Either my theory is wrong or… Well… Feynman can’t be wrong, can he? I mean… It’s not only Feynman here. We’re talking all mainstream physicists here, right?

Right. But think of it. Our electron in that thought experiment does, effectively, make a turn of 180Ā°, so it is going in the other direction now !Ā That’s more than just… Well… Going around the apparatus and looking at stuff from the other side.

Hmm… Interesting. Let’s think about the difference between theĀ sinĪø āĀ iĀ·cosĪø andĀ cosĪø āĀ iĀ·sinĪø functions. First, note that they will give us the same probabilities: the square of the absolute value of both complex numbers is the same. [It’s equal to 1 because we didn’t bother to put a coefficient in front.] Secondly, we should note that the sine and cosine functions are essentially the same. They just differ by a phase factor: cosĪø =Ā sin(Īø +Ā Ļ/2) andĀ āsinĪø =Ā cos(Īø +Ā Ļ/2). Let’s see what we can do with that. We can write the following, for example:

sinĪø āĀ iĀ·cosĪø =Ā ācos(Īø +Ā Ļ/2) āĀ iĀ·sin(Īø +Ā Ļ/2) =Ā ā[cos(Īø +Ā Ļ/2) +Ā iĀ·sin(Īø +Ā Ļ/2)] =Ā āeiĀ·(Īø +Ā Ļ/2)

Well… I guess that’s something at least ! The eiĀ·ĪøĀ and āeiĀ·(Īø +Ā Ļ/2)Ā functions differ by a phase shiftĀ andĀ a minus sign so… Well… That’s what it takes to reverse the direction of an electron. š Let us mull over that in the coming days. As I mentioned, these more philosophical topics are not easily exhausted. š

# The geometry of the wavefunction, electron spin and the form factor

Our previous posts showed how a simple geometric interpretation of the elementary wavefunction yielded the (Compton scattering) radius of an elementary particleāfor an electron, at least: for the proton, we only got the order of magnitude rightābut then a proton is not an elementary particle.Ā We got lots of other interesting equations as well… But… Well… When everything is said and done, it’s that equivalence between theĀ E =Ā mĀ·a2Ā·Ļ2Ā andĀ E =Ā mĀ·c2Ā relations that we… Well… We need to be moreĀ specific about it.

Indeed, I’ve been ambiguous here and thereāoscillatingĀ between various interpretations, so to speak. š In my own mind, I refer to my unanswered questions, or my ambiguous answers to them, as the form factorĀ problem.Ā So… Well… That explains the title of my post. But so… Well… I do want to be somewhat moreĀ conclusiveĀ in this post. So let’s go and see where we end up. š

To help focus our mind, let us recall the metaphor of the V-2 perpetuum mobile, as illustrated below. With permanently closed valves, the air inside the cylinder compresses and decompresses as the pistons move up and down. It provides, therefore, a restoring force. As such, it will store potential energy, just like a spring, and the motion of the pistons will also reflect that of a mass on a spring: it is described by a sinusoidal function, with the zero point at the center of each cylinder. We can, therefore, think of the moving pistons as harmonic oscillators, just like mechanical springs. Of course, instead of two cylinders with pistons, one may also think of connecting two springs with a crankshaft, but then that’s not fancy enough for me. š

At first sight, the analogy between our flywheel model of an electron and the V-twin engine seems to be complete: the 90 degree angle of ourĀ V-2 engine makes it possible to perfectly balance the pistons and we may, therefore, think of the flywheel as a (symmetric) rotating mass, whose angular momentum is given by the product of the angular frequency and the moment of inertia: L =Ā ĻĀ·I. Of course,Ā the moment of inertia (aka the angular mass) will depend on theĀ formĀ (orĀ shape) of our flywheel:

1. I = mĀ·a2Ā for a rotating pointĀ mass m or, what amounts to the same, for a circular hoop of mass m and radiusĀ rĀ =Ā a.
2. For a rotating (uniformly solid)Ā disk, we must add a 1/2 factor: IĀ =Ā mĀ·a2/2.

How can we relate those formulas to the E =Ā mĀ·a2Ā·Ļ2Ā formula? TheĀ kinetic energy that is being stored in a flywheel is equal EkineticĀ = IĀ·Ļ2/2, so that is only halfĀ of theĀ E =Ā mĀ·a2Ā·Ļ2Ā product if we substitute I forĀ I = mĀ·a2. [For a disk, we get a factor 1/4, so that’s even worse!] However, our flywheel model of an electron incorporates potential energy too. In fact, theĀ E =Ā mĀ·a2Ā·Ļ2Ā formula just adds the (kinetic and potential) energy of two oscillators: we do not really consider the energy in the flywheel itself because… Well… The essence of our flywheel model of an electron is not the flywheel: the flywheel justĀ transfersĀ energy from one oscillator to the other, but so… Well… We don’tĀ includeĀ it in our energy calculations. The essence of our model is thatĀ two-dimensional oscillation whichĀ drivesĀ the electron, and which is reflected in Einstein’sĀ E =Ā mĀ·c2Ā formula.Ā That two-dimensional oscillationātheĀ a2Ā·Ļ2Ā = c2Ā equation, reallyātells us that theĀ resonantĀ (orĀ natural) frequencyĀ of the fabric of spacetime is given by theĀ speed of lightābut measured in units ofĀ a. [If you don’t quite get this, re-write theĀ a2Ā·Ļ2Ā = c2Ā equation asĀ Ļ = c/a: the radius of our electron appears as a naturalĀ distance unit here.]

Now, we were extremely happy with this interpretation not only because of the key results mentioned above, but also because it has lots of other nice consequences. Think of our probabilities as being proportional to energy densities, for exampleāand all of the other stuff I describe in my published paper on this. But there is even more on the horizon: a follower of this blog (a reader with an actual PhD in physics, for a change) sent me an article analyzing elementary particles as tiny black holes because… Well… If our electron is effectively spinning around, then its tangential velocity is equal toĀ vĀ =Ā aĀ·ĻĀ =Ā c. Now, recent research suggest black holes are also spinning at (nearly) the speed of light. Interesting, right? However, in order to understand what she’s trying to tell me, I’ll first need to get a better grasp of general relativity, so I can relate what I’ve been writing here and in previous posts to the Schwarzschild radiusĀ and other stuff.

Let me get back to the lesson here. In the reference frame of our particle, the wavefunction really looks like the animation below: it has two components, and the amplitude of the two-dimensional oscillation is equal to a, which we calculated asĀ aĀ =Ā Ä§Ā·/(mĀ·c) = 3.8616Ć10ā13Ā m, so that’s the (reduced) Compton scattering radius of an electron.

In my original article on this, I used a more complicated argument involving the angular momentum formula, but I now prefer a more straightforward calculation:

cĀ = aĀ·ĻĀ =Ā aĀ·E/Ä§ =Ā aĀ·mĀ·c2/Ä§Ā Ā āĀ aĀ =Ā Ä§/(mĀ·c)

The question is: whatĀ is that rotating arrow? I’ve been vague and not so vague on this. The thing is: I can’tĀ proveĀ anything in this regard. But myĀ hypothesisĀ is that it is, in effect, aĀ rotatingĀ field vector, so it’s just like the electric field vector of a (circularly polarized) electromagnetic wave (illustrated below).

There are a number of crucial differences though:

1. The (physical) dimension of the field vector of the matter-wave is different: I associate the real and imaginary component of the wavefunction with a force per unit massĀ (as opposed to the force per unit charge dimension of the electric field vector). Of course, the newton/kg dimension reduces to the dimension of acceleration (m/s2), so that’s the dimension of a gravitational field.
2. I do believe this gravitational disturbance, so to speak, does cause an electron to move about some center, and I believe it does so at the speed of light. In contrast, electromagnetic waves doĀ notĀ involve any mass: they’re just an oscillating field. Nothing more. Nothing less. In contrast, as Feynman puts it: “When you do find the electron some place, the entire charge is there.” (Feynman’s Lectures, III-21-4)
3. The third difference is one that I thought of only recently: theĀ planeĀ of the oscillation cannotĀ be perpendicular to the direction of motion of our electron, because then we can’t explain the direction of its magnetic moment, which is either up or down when traveling through a Stern-Gerlach apparatus.

I mentioned that in my previous post but, for your convenience, I’ll repeat what I wrote there.Ā The basic idea here is illustrated below (credit for this illustration goes toĀ another blogger on physics). As for the Stern-Gerlach experiment itself, let me refer you to aĀ YouTube videoĀ from theĀ Quantum Made SimpleĀ site.

The point is: the direction of the angular momentum (and the magnetic moment) of an electronāor, to be precise, its component as measured in the direction of the (inhomogeneous) magnetic field through which our electron isĀ travelingācannotĀ be parallel to the direction of motion. On the contrary, it isĀ perpendicularĀ to the direction of motion. In other words, if we imagine our electron as spinning around some center, then the disk it circumscribes will compriseĀ the direction of motion.

However, we need to add an interesting detail here. As you know, we don’t really have a precise direction of angular momentum in quantum physics. [If you don’t know this… Well… Just look at one of my many posts on spin and angular momentum in quantum physics.] Now, we’ve explored a number of hypotheses but, when everything is said and done, a rather classical explanation turns out to be the best: an object with an angular momentum JĀ and a magnetic momentĀ Ī¼Ā (I used bold-face because these areĀ vector quantities) that is parallel to some magnetic field B, will notĀ line up, as you’d expect a tiny magnet to do in a magnetic fieldāor not completely, at least: it willĀ precess. I explained that in another post on quantum-mechanical spin, which I advise you to re-read if you want to appreciate the point that I am trying to make here. That post integrates some interesting formulas, and so one of the things on my ‘to do’ list is to prove that these formulas are, effectively, compatible with the electron model we’ve presented in this and previous posts.

Indeed, when one advances a hypothesis like this, it’s not enough to just sort ofĀ showĀ that the general geometry of the situation makes sense: we also need to show the numbers come out alright. So… Well… Whatever weĀ thinkĀ our electronāor its wavefunctionāmight be, it needs to be compatible with stuff like the observedĀ precession frequencyĀ of an electron in a magnetic field.

Our model also needs to be compatible with the transformation formulas for amplitudes. I’ve been talking about this for quite a while now, and so it’s about time I get going on that.

Last but not least, those articles that relate matter-particles to (quantum) gravityāsuch as the one I mentioned aboveāare intriguing too and, hence, whatever hypotheses I advance here, I’d better check them against those more advanced theories too, right? š Unfortunately, that’s going to take me a few more years of studying… But… Well… I still have many years aheadāI hope. š

Post scriptum: It’s funny how one’s brain keeps working when sleeping. When I woke up this morning, I thought: “But itĀ isĀ that flywheel that matters, right? That’s the energy storage mechanism and also explains how photons possibly interact with electrons. The oscillatorsĀ driveĀ the flywheel but, without the flywheel, nothing is happening. It is really theĀ transferĀ of energyāthrough the flywheelāwhich explains why our flywheel goes round and round.”

It may or may not be useful to remind ourselves of the math in this regard.Ā The motionĀ ofĀ our first oscillator is given by the cos(ĻĀ·t) = cosĪø function (Īø = ĻĀ·t), and its kinetic energy will be equal toĀ sin2Īø. Hence, the (instantaneous)Ā changeĀ in kinetic energy at any point in time (as a function of the angle Īø) isĀ equal to:Ā d(sin2Īø)/dĪø = 2āsinĪøād(sinĪø)/dĪø = 2āsinĪøācosĪø. Now, the motion of theĀ second oscillator (just look at that second piston going up and down in the V-2 engine) is given by theĀ sinĪø function, which is equal to cos(Īø ā Ļ /2). Hence, its kinetic energy is equal toĀ sin2(Īø ā Ļ /2), and how itĀ changesĀ (as a function of Īø again) is equal toĀ 2āsin(Īø ā Ļ /2)ācos(Īø ā Ļ /2) =Ā = ā2ācosĪøāsinĪø = ā2āsinĪøācosĪø. So here we have our energy transfer: the flywheel organizes the borrowing and returning of energy, so to speak. That’s the crux of the matter.

So… Well… WhatĀ if the relevant energy formula isĀ E =Ā mĀ·a2Ā·Ļ2/2 instead ofĀ E =Ā mĀ·a2Ā·Ļ2? What are the implications? Well… We get aĀ ā2 factor in our formula for the radiusĀ a, as shown below.

Now that isĀ notĀ so nice. For the tangential velocity, we getĀ vĀ =Ā aĀ·Ļ =Ā ā2Ā·c. This is alsoĀ notĀ so nice. How can we save our model? I am not sure, but here I am thinking of the mentioned precessionātheĀ wobbling of our flywheel in a magnetic field. Remember we may think of Jzāthe angular momentum or, to be precise, its component in theĀ z-direction (the direction in which weĀ measureĀ itāas the projection of theĀ realĀ angular momentumĀ J. Let me insert Feynman’s illustration here again (Feynman’s Lectures, II-34-3), so you get what I am talking about.

Now, all depends on the angle (Īø) betweenĀ JzĀ andĀ J, of course. We did a rather obscure post on these angles, but the formulas there come in handy now. Just click the link and review it if and when you’d want to understand the following formulas for theĀ magnitudeĀ of theĀ presumedĀ actualĀ momentum:In this particular case (spin-1/2 particles),Ā j is equal to 1/2 (in units ofĀ Ä§, of course). Hence,Ā JĀ is equal toĀ ā0.75Ā ā 0.866. Elementary geometry then tells us cos(Īø) =Ā (1/2)/ā(3/4) =Ā  = 1/ā3. Hence,Ā ĪøĀ ā 54.73561Ā°. That’s a big angleālarger than the 45Ā° angle we had secretly expected because… Well… The 45Ā° angle has thatĀ ā2 factor in it:Ā cos(45Ā°) =Ā sin(45Ā°) = 1/ā2.

Hmm… As you can see, there is no easy fix here. Those damn 1/2 factors! They pop up everywhere, don’t they? š We’ll solve the puzzle. One day… But not today, I am afraid. I’ll call it the form factor problem… Because… Well… It sounds better than the 1/2 orĀ ā2 problem, right? š

Note: If you’re into quantum math, you’ll noteĀ aĀ =Ā Ä§/(mĀ·c) is theĀ reducedĀ Compton scattering radius. The standard Compton scattering radius is equal toĀ aĀ·2ĻĀ = (2ĻĀ·Ä§)/(mĀ·c) =Ā  h/(mĀ·c) = h/(mĀ·c). It doesn’t solve theĀ ā2 problem. Sorry. The form factor problem. š

To be honest, I finished my published paper on all of this with a suggestion that, perhaps, we should think of twoĀ circularĀ oscillations, as opposed to linear ones. Think of a tiny ball, whose center of mass stays where it is, as depicted below. Any rotation ā around any axis ā will be some combination of a rotation around the two other axes. Hence, we may want to think of our two-dimensionalĀ oscillation as an oscillation of a polar and azimuthal angle. It’s just a thought but… Well… I am sure it’s going to keep me busy for a while. šThey are oscillations, still, so I am not thinking ofĀ twoĀ flywheels that keep going around in the same direction. No. More like a wobbling object on a spring. Something like the movement of a bobblehead on a spring perhaps. š

# The speed of light as an angular velocity

Over the weekend, I worked on a revised version of my paper on a physical interpretation of the wavefunction. However, I forgot to add the final remarks on the speed of light as an angular velocity. I know… This post is for my faithful followers only. It is dense, but let me add the missing bits here:

Post scriptum (29 October):Ā Einsteinās view on aether theories probably still holds true: āWe may say that according to the general theory of relativity space is endowed with physical qualities; in this sense, therefore, there exists an aether. According to the general theory of relativity, space without aether is unthinkable ā for in such space there not only would be no propagation of light, but also no possibility of existence for standards of space and time (measuring-rods and clocks), nor therefore any space-time intervals in the physical sense. But this aether may not be thought of as endowed with the quality characteristic of ponderable media, as consisting of parts which may be tracked through time. The idea of motion may not be applied to it.ā

The above quote is taken from the Wikipedia article on aether theories. The same article also quotes Robert Laughlin, the 1998 Nobel Laureate in Physics, who said this about aether in 2005: āIt is ironic that Einstein’s most creative work, the general theory of relativity, should boil down to conceptualizing space as a medium when his original premise [in special relativity] was that no such medium existed. [ā¦] The word ‘aether’ has extremely negative connotations in theoretical physics because of its past association with opposition to relativity. This is unfortunate because, stripped of these connotations, it rather nicely captures the way most physicists actually think about the vacuum. [ā¦]The modern concept of the vacuum of space, confirmed every day by experiment, is a relativistic aether. But we do not call it this because it is taboo.ā

I really love this: a relativistic aether. MyĀ interpretation of the wavefunction is veryĀ consistent with that.

# A physical explanation for relativistic length contraction?

My last posts were all about a possible physicalĀ interpretation of the quantum-mechanical wavefunction. To be precise, we have been interpreting the wavefunction as a gravitational wave. In this interpretation, the real and imaginary component of the wavefunction get aĀ physicalĀ dimension: force per unit mass (newton per kg). The inspiration here was the structural similarity between Coulomb’s and Newton’s force laws. They both look alike: it’s just that one gives us a force per unitĀ chargeĀ (newton perĀ coulomb), while the other gives us a force per unitĀ mass.

So… Well… Many nice things came out of this – and I wrote about that at length – but last night I was thinking this interpretation may also offer an explanation of relativistic length contraction. Before we get there, let us re-visit our hypothesis.

The geometry of the wavefunction

The elementary wavefunction is written as:

Ļ =Ā aĀ·eāi(EĀ·t ā pāx)/Ä§ =Ā aĀ·cos(pāx/Ä§ – Eāt/Ä§) + iĀ·aĀ·sin(pāx/Ä§ – Eāt/Ä§)

Nature should not care about our conventions for measuring the phase angle clockwise or counterclockwise and, therefore, the Ļ =Ā aĀ·ei[EĀ·t ā pāx]/Ä§ function may also be permitted. We know that cos(Īø) = cos(Īø) and sinĪø = sin(Īø), so we can write:Ā  Ā Ā

Ļ =Ā aĀ·ei(EĀ·t ā pāx)/Ä§ =Ā aĀ·cos(Eāt/Ä§ – pāx/Ä§) + iĀ·aĀ·sin(Eāt/Ä§ – pāx/Ä§)

= aĀ·cos(pāx/Ä§ – Eāt/Ä§) iĀ·aĀ·sin(pāx/Ä§ – Eāt/Ä§)

The vectors p and x are the the momentum and position vector respectively: p = (px, py, pz) and x = (x, y, z). However, if we assume there is no uncertainty about p ā not about the direction nor the magnitude ā then we may choose an x-axis which reflects the direction of p. As such, x = (x, y, z) reduces to (x, 0, 0), and pāx/Ä§ reduces to pāx/Ä§. This amounts to saying our particle is traveling along the x-axis or, if p = 0, that our particle is located somewhere on the x-axis. Hence, the analysis is one-dimensional only.

The geometry of the elementary wavefunction is illustrated below. The x-axis is the direction of propagation, and the y- and z-axes represent the real and imaginary part of the wavefunction respectively.

Note that, when applying the right-hand rule for the axes, the vertical axis is the y-axis, not the z-axis. Hence, we may associate the vertical axis with the cosine component, and the horizontal axis with the sine component. You can check this as follows: if the origin is the (x, t) = (0, 0) point, then cos(Īø) = cos(0) = 1 and sin(Īø) = sin(0) = 0. This is reflected in both illustrations, which show a left- and a right-handed wave respectively. We speculated this should correspond to the two possible values for the quantum-mechanical spin of the wave: +Ä§/2 or āÄ§/2. The cosine and sine components for the left-handed wave are shown below. Needless to say, the cosine and sine function are the same, except for a phase difference of Ļ/2: sin(Īø) = cos(Īø ā Ļ/2).

As for the wave velocity, and its direction of propagation, we know that the (phase) velocity of any wave F(kx – Ļt) is given by vp =Ā Ļ/k = (E/Ä§)/(p/Ä§) = E/p. Of course, the momentum might also be in the negative x-direction, in which case k would be equal to -p and, therefore, we would get a negative phase velocity: vp =Ā Ļ/k = E/p.

The de Broglie relations

E/Ä§ = Ļ gives the frequency in time (expressed in radians per second), while p/Ä§ = k gives us the wavenumber, or the frequency in space (expressed in radians per meter). Of course, we may write: f = Ļ/2ĻĀ  and Ī» = 2Ļ/k, which gives us the two de Broglie relations:

1. E = Ä§āĻ = hāf
2. p = Ä§āk = h/Ī»

The frequency in time is easy to interpret. The wavefunction of a particle with more energy, or more mass, will have a higher density in time than a particle with less energy.

In contrast, the second de Broglie relation is somewhat harder to interpret. According to the p = h/Ī» relation, the wavelength is inversely proportional to the momentum: Ī» = h/p. The velocity of a photon, or a (theoretical) particle with zero rest mass (m0 = 0), is c and, therefore, we find that p = mvāv = mcāc = māc (all of the energy is kinetic). Hence, we can write: pāc = māc2 = E, which we may also write as: E/p = c. Hence, for a particle with zero rest mass, the wavelength can be written as:

Ī» = h/p = hc/E = h/mc

However, this is a limiting situation ā applicable to photons only. Real-life matter-particles should have some mass[1] and, therefore, their velocity will never be c.[2]

Hence, if p goes to zero, then the wavelength becomes infinitely long: if pĀ ā 0 then Ī» āĀ ā. How should we interpret this inverse proportionality between Ī» and p? To answer this question, let us first see what this wavelength Ī» actually represents.

If we look at the Ļ = aĀ·cos(pāx/Ä§ – Eāt/Ä§) – iĀ·aĀ·sin(pāx/Ä§ – Eāt/Ä§) once more, and if we write pāx/Ä§ as Ī, then we can look at pāx/Ä§ as a phase factor, and so we will be interested to know for what x this phase factor Ī = pāx/Ä§ will be equal to 2Ļ. So we write:

Ī =pāx/Ä§ = 2Ļ āĀ x = 2ĻāÄ§/p = h/p = Ī»

So now we get a meaningful interpretation for that wavelength. It is the distance between the crests (or the troughs) of the wave, so to speak, as illustrated below. Of course, this two-dimensional wave has no real crests or troughs: we measure crests and troughs against the y-axis here. Hence, our definition depend on the frame of reference.

Now we know what Ī» actually represents for our one-dimensional elementary wavefunction. Now, the time that is needed for one cycle is equal to T = 1/f = 2ĻĀ·(Ä§/E). Hence, we can now calculate the wave velocity:

v = Ī»/T = (h/p)/[2ĻĀ·(Ä§/E)] = E/p

Unsurprisingly, we just get the phase velocity that we had calculated already: v = vp = E/p. The question remains: what if p is zero? What if we are looking at some particle at rest? It is an intriguing question: we get an infinitely long wavelength, and an infinite wave velocity.

Now, re-writing the v = E/p as v = māc2/māvg Ā = c/Ī²g, in which Ī²g is the relative classical velocity[3] of our particle Ī²g = vg/c) tells us that the phase velocities will effectively be superluminal (Ī²gĀ  < 1 so 1/ Ī²g > 1), but what if Ī²g approaches zero? The conclusion seems unavoidable: for a particle at rest, we only have a frequency in time, as the wavefunction reduces to:

Ļ =Ā aĀ·eāiĀ·EĀ·t/Ä§ =Ā aĀ·cos(Eāt/Ä§) – iĀ·aĀ·sin(Eāt/Ä§)

How should we interpret this?

A physical interpretation of relativistic length contraction?

In my previous posts,Ā we argued that the oscillations of the wavefunction pack energy. Because the energy of our particle is finite, the wave train cannot be infinitely long. If we assume some definite number of oscillations, then the string of oscillations will be shorter as Ī» decreases. Hence, the physical interpretation of the wavefunction that is offered here may explain relativistic length contraction.

š

[1] Even neutrinos have some (rest) mass. This was first confirmed by the US-Japan Super-Kamiokande collaboration in 1998. Neutrinos oscillate between three so-called flavors: electron neutrinos, muon neutrinos and tau neutrinos. Recent data suggests that the sum of their masses is less than a millionth of the rest mass of an electron. Hence, they propagate at speeds that are very near to the speed of light.

[2] Using the Lorentz factor (Ī³), we can write the relativistically correct formula for the kinetic energy as KE = EĀ āĀ E0Ā =Ā mvc2Ā ā m0c2Ā =Ā m0Ī³c2Ā ā m0c2Ā =Ā m0c2(Ī³ ā 1). As v approaches c, Ī³ approaches infinity and, therefore, the kinetic energy would become infinite as well.

[3] Because our particle will be represented by a wave packet, i.e. a superimposition of elementary waves with different E and p, the classical velocity of the particle becomes the group velocity of the wave, which is why we denote it by vg.

# The geometry of the wavefunction (2)

This post further builds on the rather remarkable results we got in our previous posts. Let us start with the basics once again.Ā The elementary wavefunction is written as:

Ļ =Ā aĀ·eāi[EĀ·tĀ ā pāx]/Ä§ =Ā aĀ·cos(pāx/Ä§Ā ā Eāt/Ä§) + iĀ·aĀ·sin(pāx/Ä§Ā ā Eāt/Ä§)

Of course, NatureĀ (or God, as Einstein would put it) does not care about our conventions for measuring an angle (i.e. the phase of our wavefunction) clockwise or counterclockwise and, therefore, the Ļ =Ā aĀ·ei[EĀ·t ā pāx]/Ä§ function is also permitted. We know that cos(Īø) = cos(āĪø) and sinĪø = āsin(āĪø), so we can write:Ā  Ā Ā

Ļ =Ā aĀ·ei[EĀ·t ā pāx]/Ä§ =Ā aĀ·cos(Eāt/Ä§Ā āĀ pāx/Ä§) + iĀ·aĀ·sin(Eāt/Ä§Ā āĀ pāx/Ä§)

= aĀ·cos(pāx/Ä§Ā ā Eāt/Ä§) āĀ iĀ·aĀ·sin(pāx/Ä§Ā ā Eāt/Ä§)

The vectors p and x are the momentum and position vector respectively: p = (px, py, pz) and x = (x, y, z). However, if we assume there is no uncertainty about p ā not about the direction, and not about the magnitude ā then the direction of p can be our x-axis. In this reference frame,Ā x = (x, y, z) reduces to (x, 0, 0), and pāx/Ä§ reduces to pāx/Ä§. This amounts to saying our particle is traveling along the x-axis or, if p = 0, that our particle is located somewhere on the x-axis. So we have an analysis in one dimension only then, which facilitates our calculations. The geometry of the wavefunction is then as illustrated below. The x-axis is the direction of propagation, and the y- and z-axes represent the real and imaginary part of the wavefunction respectively.

Note that, when applying the right-hand rule for the axes, the vertical axis is the y-axis, not the z-axis. Hence, we may associate the vertical axis with the cosine component, and the horizontal axis with the sine component. [You can check this as follows: if the origin is the (x, t) = (0, 0) point, then cos(Īø) = cos(0) = 1 and sin(Īø) = sin(0) = 0. This is reflected in both illustrations, which show a left- and a right-handed wave respectively.]

Now, you will remember that we speculated the two polarizations (left- versus right-handed) should correspond to the two possible values for the quantum-mechanical spin of the wave (+Ä§/2 or āÄ§/2). We will come back to this at the end of this post. Just focus on the essentials first: the cosine and sine components for the left-handed wave are shown below. Look at it carefully and try to understand. Needless to say, the cosine and sine function are the same, except for a phase difference of Ļ/2: sin(Īø) = cos(Īø ā Ļ/2).

As for the wave velocity, and its direction of propagation, we know that the (phase) velocity of any waveform F(kx ā Ļt) is given by vp =Ā Ļ/k. In our case, we find thatĀ vp =Ā Ļ/k = (E/Ä§)/(p/Ä§) = E/p. Of course, the momentum might also be in the negative x-direction, in which case k would be equal to āp and, therefore, we would get a negative phase velocity: vp =Ā Ļ/k = (E/Ä§)/(āp/Ä§) = āE/p.

As you know, E/Ä§ = Ļ gives the frequency in time (expressed in radians per second), while p/Ä§ = k gives us the wavenumber, or the frequency in space (expressed in radians per meter). [If in doubt, check my post on essential wave math.] Now, you also know that f = Ļ/2ĻĀ  and Ī» = 2Ļ/k, which gives us the two de Broglie relations:

1. E = Ä§āĻ = hāf
2. p = Ä§āk = h/Ī»

The frequency in time (oscillations or radians per second) is easy to interpret. A particle will always have some mass and, therefore, some energy, and it is easy to appreciate the fact that the wavefunction of a particle with more energy (or more mass) will have a higher density in time than a particle with less energy.

However, the second de Broglie relation is somewhat harder to interpret. Note that the wavelength is inversely proportional to the momentum: Ī» = h/p. Hence, if p goes to zero, then the wavelength becomes infinitely long, so we write:

If pĀ ā 0 then Ī» āĀ ā.

For the limit situation, a particle with zero rest mass (m0 = 0), the velocity may be c and, therefore, we find that p = mvāv = mcāc = māc (all of the energy is kinetic) and, therefore, pāc = māc2 = E, which we may also write as: E/p = c. Hence, for a particle with zero rest mass (m0Ā = 0), the wavelength can be written as:

Ī» = h/p = hc/E = h/mc

Of course, we are talking a photon here. We get the zero rest mass for a photon. In contrast, all matter-particles should have some mass[1] and, therefore, their velocity will neverĀ equalĀ c.[2] The question remains: how should we interpret the inverse proportionality between Ī» and p?

Let us first see what this wavelength Ī» actually represents. If we look at the Ļ = aĀ·cos(pāx/Ä§ ā Eāt/Ä§) āĀ iĀ·aĀ·sin(pāx/Ä§ – Eāt/Ä§) once more, and if we write pāx/Ä§ as Ī, then we can look at pāx/Ä§ as a phase factor, and so we will be interested to know for what x this phase factor Ī = pāx/Ä§ will be equal to 2Ļ. So we write:

Ī =pāx/Ä§ = 2Ļ āĀ x = 2ĻāÄ§/p = h/p = Ī»

So now we get a meaningful interpretation for that wavelength. It is the distance between the crests (or the troughs) of the wave, so to speak, as illustrated below. Of course, this two-dimensional wave has no real crests or troughs: they depend on your frame of reference.

So now we know what Ī» actually represent for our one-dimensional elementary wavefunction. Now, the time that is needed for one cycle is equal to T = 1/f = 2ĻĀ·(Ä§/E). Hence, we can now calculate the wave velocity:

v = Ī»/T = (h/p)/[2ĻĀ·(Ä§/E)] = E/p

Unsurprisingly, we just get the phase velocity that we had calculated already: v = vp = E/p. It does not answer the question: what if p is zero? What if we are looking at some particle at rest? It is an intriguing question: we get an infinitely long wavelength, and an infinite phase velocity. Now, we know phase velocities can be superluminal, but they should not be infinite. So what does the mathematical inconsistency tell us? Do these infinitely long wavelengths and infinite wave velocities tell us that our particle has to move? Do they tell us our notion of a particle at rest is mathematically inconsistent?

Maybe. But maybe not. Perhaps the inconsistency just tells us our elementary wavefunction ā or the concept of a precise energy, and a precise momentum ā does not make sense. This is where the Uncertainty Principle comes in: stating that p = 0, implies zero uncertainty. Hence, the Ļp factor in the ĻpāĻxĀ ā¤ Ä§/2 would be zero and, therefore, ĻpāĻx would be zero which, according to the Uncertainty Principle, it cannot be: it can be very small, but it cannot be zero.

It is interesting to note here that Ļp refers to the standard deviation from the mean, as illustrated below. Of course, the distribution may be or may not be normal ā we donāt know ā but a normal distribution makes intuitive sense, of course. Also, if we assume the mean is zero, then the uncertainty is basically about the direction in which our particle is moving, as the momentum might then be positive or negative.

The question of natural units may pop up. The Uncertainty Principle suggests a numerical value of the natural unit for momentum and distance that is equal to the square root of Ä§/2, so thatās about 0.726Ć10ā17 m for the distance unit and 0.726Ć10ā17 Nās for the momentum unit, as the product of both gives us Ä§/2. To make this somewhat more real, we may note that 0.726Ć10ā17 m is the attometer scale (1 am = 1Ć10ā18 m), so that is very small but not unreasonably small.[3]

Hence, we need to superimpose a potentially infinite number of waves with energies and momenta centered on some mean value. It is only then that we get meaningful results. For example, the idea of a group velocity ā which should correspond to the classical idea of the velocity of our particle ā only makes sense in the context of wave packet. Indeed, the group velocity of a wave packet (vg) is calculated as follows:

vgĀ = āĻi/ākiĀ = ā(Ei/Ä§)/ā(pi/Ä§) = ā(Ei)/ā(pi)

This assumes the existence of a dispersion relation which gives us ĻiĀ as a function of ki ā what amounts to the same ā EiĀ as a function of pi. How do we get that?Ā Well… There are a few ways to go about it but one interesting way of doing it is to re-write SchrĆ¶dinger’s equation as the following pairĀ of equations[4]:

1. Re(āĻ/āt) = ā[Ä§/(2meff)]Ā·Im(ā2Ļ)Ā āĀ ĻĀ·cos(kx ā Ļt) =Ā k2Ā·[Ä§/(2meff)]Ā·cos(kx ā Ļt)
2. Im(āĻ/āt) = [Ä§/(2meff)]Ā·Re(ā2Ļ)Ā ā ĻĀ·sin(kx ā Ļt) = k2Ā·[Ä§/(2meff)]Ā·sin(kx ā Ļt)

These equations imply the following dispersion relation:

Ļ =Ā Ä§Ā·k2/(2m)

Of course, we need to think about the subscripts now: we have Ļi, ki, but… What about meff or, dropping the subscript, about m? Do we write it as mi? If so, what is it? Well… It is the equivalent mass of EiĀ obviously, and so we get it from the mass-energy equivalence relation: miĀ = Ei/c2. It is a fine point, but one most people forget about: they usually just write m. However, if there is uncertainty in the energy, then Einstein’s mass-energy relation tells us we must have some uncertainty in the (equivalent) mass too, and the two will, obviously, be related as follows: Ļm = ĻE/c2. We are tempted to do a few substitutions here. Letās first check what we get when doing the miĀ = Ei/c2 substitution:

Ļi =Ā Ä§Ā·ki2/(2mi) = (1/2)āÄ§Ā·ki2āc2/Ei = (1/2)āÄ§Ā·ki2āc2/(ĻiāÄ§)Ā = (1/2)āÄ§Ā·ki2āc2/Ļi

ā Ļi2/ki2 = c2/2Ā ā Ļi/ki = vp = c/2 !?

We get a very interesting but nonsensical condition for the dispersion relation here. I wonder what mistake I made. š¦

Let us try another substitution. The group velocity is what it is, right? It is the velocity of the group, so we can write: ki = p/Ä§ = mi Ā·vg. This gives us the following result:

Ļi =Ā Ä§Ā·(mi Ā·vg)2/(2mi) = Ä§Ā·miĀ·vg2/2

It is yet another interesting condition for the dispersion relation. Does it make any more sense? I am not so sure. That factor 1/2 troubles us. It only makes sense when weĀ dropĀ it. Now you will object that SchrĆ¶dinger’s equation gives us the electron orbitals – and many other correct descriptions of quantum-mechanical stuff – so, surely,Ā SchrĆ¶dinger’s equation cannot be wrong. You’re right. It’s just that… Well… When we are splitting in up in two equations, as we are doing here, then we are looking atĀ oneĀ of the two dimensions of the oscillation only and, therefore, it’s onlyĀ halfĀ of the mass that counts. Complicated explanation but… Well… It should make sense, because the results that come out make sense. Think of it. So we write this:

• Re(āĻ/āt) = ā(Ä§/meff)Ā·Im(ā2Ļ)Ā āĀ ĻĀ·cos(kx ā Ļt) =Ā k2Ā·(Ä§/meff)Ā·cos(kx ā Ļt)
• Im(āĻ/āt) = (Ä§/meff)Ā·Re(ā2Ļ)Ā ā ĻĀ·sin(kx ā Ļt) = k2Ā·(Ä§/meff)Ā·sin(kx ā Ļt)

We then get the dispersion relation withoutĀ that 1/2 factor:

Ļi =Ā Ä§Ā·ki2/mi

TheĀ miĀ = Ei/c2 substitution then gives us the result we sort of expected to see:

Ļi =Ā Ä§Ā·ki2/miĀ = Ä§Ā·ki2āc2/Ei = Ä§Ā·ki2āc2/(ĻiāÄ§)Ā ā Ļi/ki = vp = c

Likewise, the other calculation also looks more meaningful now:

Ļi =Ā Ä§Ā·(mi Ā·vg)2/miĀ = Ä§Ā·miĀ·vg2

Sweet ! š

Let us put this aside for the moment and focus on something else. If you look at the illustrations above, you see we can sort of distinguish (1) a linear velocity ā the speed with which those wave crests (or troughs) move ā and (2) some kind of circular or tangential velocity ā the velocity along the red contour lineĀ above. Weāll need the formula for a tangential velocity: vt = aāĻ.

Now, if Ī» is zero, then vt = aāĻ = aāE/Ä§ is just all there is. We may double-check this as follows: the distance traveled in one period will be equal to 2Ļa, and the period of the oscillation is T = 2ĻĀ·(Ä§/E). Therefore, vt will, effectively, be equal to vt = 2Ļa/(2ĻÄ§/E) = aāE/Ä§.Ā However, if Ī» is non-zero, then the distance traveled in one period will be equal to 2Ļa + Ī». The period remains the same: T = 2ĻĀ·(Ä§/E). Hence, we can write:

For an electron, we did this weird calculation. We had an angular momentum formula (for an electron) which we equated with theĀ real-life +Ä§/2 or āÄ§/2 values of its spin, and we got aĀ numericalĀ value forĀ a. It was the Compton radius: the scattering radius for an electron. Let us write it out:

Using the right numbers, youāll find the numerical value for a: 3.8616Ć10ā13 m. But let us just substitute the formula itself here:Ā

This is fascinating ! And we just calculated that vpĀ is equal toĀ c. For the elementary wavefunction, that is. Hence, we get this amazing result:

vt = 2c

ThisĀ tangentialĀ velocity isĀ twiceĀ the linearĀ velocity !

Of course, the question is: what is theĀ physicalĀ significance of this? I need to further look at this. Wave velocities are, essentially, mathematicalĀ concepts only: the wave propagates through space, butĀ nothing elseĀ is really moving. However, the geometric implications are obviously quite interesting and, hence, need further exploration.

One conclusion stands out: all these results reinforce our interpretation of the speed of light as aĀ propertyĀ of the vacuum – or of the fabric of spacetime itself. š

[1] Even neutrinos should have some (rest) mass. In fact, the mass of the known neutrino flavors was estimated to be smaller than 0.12 eV/c2. This mass combines the three known neutrino flavors.

[2] Using the Lorentz factor (Ī³), we can write the relativistically correct formula for the kinetic energy as KE = EĀ āĀ E0Ā =Ā mvc2Ā ā m0c2Ā =Ā m0Ī³c2Ā ā m0c2Ā =Ā m0c2(Ī³ ā 1). As v approaches c, Ī³ approaches infinity and, therefore, the kinetic energy would become infinite as well.

[3] It is, of course, extremely small, but 1 am is the current sensitivity of the LIGO detector for gravitational waves. It is also thought of as the upper limit for the length of an electron, for quarks, and for fundamental strings in string theory. It is, in any case, 1,000,000,000,000,000,000 times larger than the order of magnitude of the Planck length (1.616229(38)Ć10ā35 m).

[4] The meff is the effective mass of the particle, which depends on the medium. For example, an electron traveling in a solid (a transistor, for example) will have a different effective mass than in an atom. In free space, we can drop the subscript and just write meff = m. As for the equations, they are easily derived from noting that two complex numbers a +Ā iāb and c +Ā iād are equal if, and only if, their real and imaginary parts are the same. Now, the āĻ/āt =Ā iā(Ä§/meff)āā2Ļ equation amounts to writing something like this: a +Ā iāb =Ā iā(c +Ā iād). Now, remembering thatĀ i2Ā = ā1, you can easily figure out thatĀ iā(c +Ā iād) =Ā iāc +Ā i2ād = ā d +Ā iāc.

# The geometry of the wavefunction

My posts and article on the wavefunction as a gravitational wave are rather short on the exact geometry of the wavefunction, so let us explore that a bit here. By now, you know the formula for theĀ elementary wavefunction by heart:

Ļ =Ā aĀ·eāi[EĀ·t ā pāx]/Ä§ =Ā aĀ·cos(pāx/Ä§ ā Eāt/Ä§) + iĀ·aĀ·sin(pāx/Ä§ ā Eāt/Ä§)

If we assume the momentum p is all in the x-direction, then the p and x vectors will have the same direction, and pāx/Ä§ reduces to pāx/Ä§. This amounts to saying our particle is traveling along the x-axis. The geometry of the wavefunction is illustrated below. The x-axis is the direction of propagation, and the y- and z-axes represent the real and imaginary part of the wavefunction respectively.

Note that, when applying the right-hand rule for the axes, the vertical axis is the y-axis, not the z-axis. Hence, we may associate the vertical axis with the cosine component, and the horizontal axis with the sine component. If the origin is the (x, t) = (0, 0) point, then cos(Īø) = cos(0) = 1 and sin(Īø) = sin(0) = 0. This is reflected in both illustrations, which show a left- and a right-handed wave respectively. I am convinced these correspond to the two possible values for the quantum-mechanical spin of the wave: +Ä§/2 or āÄ§/2. But… Well… Who am I? The cosine and sine components are shown below. Needless to say, the cosine and sine function are the same, except for a phase difference of Ļ/2: sin(Īø) = cos(Īø ā Ļ/2) Ā

Surely, Nature doesn’t care a hoot about our conventions for measuring the phase angle clockwise or counterclockwise and therefore, the Ļ =Ā aĀ·ei[EĀ·t ā pāx]/Ä§ function should, effectively, also be permitted. We know that cos(Īø) = cos(Īø) and sinĪø = sin(Īø), so we can write: Ā Ā Ā

Ļ =Ā aĀ·ei[EĀ·t ā pāx]/Ä§ =Ā aĀ·cos(Eāt/Ä§ ā pāx/Ä§) + iĀ·aĀ·sin(Eāt/Ä§ ā pāx/Ä§)

= aĀ·cos(pāx/Ä§ ā Eāt/Ä§) āĀ iĀ·aĀ·sin(pāx/Ä§ ā Eāt/Ä§)

E/Ä§ = Ļ gives the frequency in time (expressed in radians per second), while p/Ä§ = k gives us the wavenumber, or the frequency in space (expressed in radians per meter). Of course, we may write: f = Ļ/2ĻĀ  and Ī» = 2Ļ/k, which gives us the two de Broglie relations:

1. E = Ä§āĻ = hāf
2. p = Ä§āk = h/Ī»

The frequency in time is easy to interpret (a particle will always have some mass and, therefore, some energy), but the wavelength is inversely proportional to the momentum: Ī» = h/p. Hence, if p goes to zero, then the wavelength becomes infinitely long: if pĀ ā 0, then Ī» āĀ ā.Ā For the limit situation, a particle with zero rest mass (m0 = 0), the velocity may be c and, therefore, we find that p = mvāv = māc Ā and, therefore, pāc = māc2 = E, which we may also write as: E/p = c. Hence, for a particle with zero rest mass, the wavelength can be written as:

Ī» = h/p = hc/E = h/mc

However, we argued that the physical dimension of the components of the wavefunction may be usefully expressed in N/kg units (force per unit mass), while the physical dimension of the electromagnetic wave are expressed in N/C (force per unit charge). This, in fact, explains the dichotomy between bosons (photons) and fermions (spin-1/2 particles). Hence, all matter-particles should have some mass.[1] But how we interpret the inverse proportionality between Ī» and p?

We should probably first ask ourselves what wavelength we are talking about. The wave only has a phase velocity here, which is equal to vp =Ā Ļ/k = (E/Ä§)/(p/Ä§) = E/p. Of course, we know that, classically, the momentum will be equal to the group velocity times the mass: p = mĀ·vg. However, when p is zero, we have a division by zero once more: if pĀ ā 0, then vp = E/pĀ ā ā. Infinite wavelengths and infinite phase velocities probably tell us that our particle has to move: our notion of a particle at rest is mathematically inconsistent. If we associate this elementary wavefunction with some particle, and if we then imagine it to move, somehow, then we get an interesting relation between the group and the phase velocity:

vpĀ =Ā Ļ/k = E/p = E/(mĀ·vg) = (mĀ·c2)/(mĀ·vg) = c2/vg

We can re-write this as vpĀ·vgĀ = c2, which reminds us of the relationship between the electric and magnetic constant (1/Īµ0)Ā·(1/Ī¼0) = c2. But what is the group velocity of the elementary wavefunction? Is it a meaningful concept?

The phase velocity is just the ratio of Ļ/k. In contrast, the group velocity is the derivative of Ļ with respect to k. So we need to write Ļ as a function of k. Can we do that even if we have only one wave? We doĀ notĀ have a wave packet here, right? Just some hypotheticalĀ building blockĀ of a real-life wavefunction, right? Right. So we should introduce uncertainty about E and p and build up the wave packet, right? Well… Yes. But let’s wait with that, and see how far we can go in our interpretation of thisĀ elementaryĀ wavefunction. Let’s first get thatĀ Ļ = Ļ(k) relation. You’ll remember we can write SchrĆ¶dingerās equation – the equation that describes theĀ propagationĀ mechanism for matter-waves – asĀ the following pair of equations:

1. Re(āĻ/āt) = ā[Ä§/(2m)]Ā·Im(ā2Ļ)Ā āĀ ĻĀ·cos(kx ā Ļt) =Ā k2Ā·[Ä§/(2m)]Ā·cos(kx ā Ļt)
2. Im(āĻ/āt) = [Ä§/(2m)]Ā·Re(ā2Ļ)Ā ā ĻĀ·sin(kx ā Ļt) = k2Ā·[Ä§/(2m)]Ā·sin(kx ā Ļt)

This tells us that Ļ =Ā Ä§Ā·k2/(2m). Therefore, we can calculate āĻ/ākĀ as:

āĻ/āk = Ä§Ā·k/m = p/m = vg

We learn nothing new. We are going round and round in circles here, and we always end up with a tautology: as soon as we have a non-zero momentum, we have a mathematicalĀ formula for the group velocity – but we don’t know what it represents – and a finite wavelength. In fact, using the p = Ä§āk = h/Ī» relation, we can write one as a function of the other:

Ī» = h/p = h/mvg āĀ vg = h/mĪ»

What does this mean? ItĀ resembles the c = h/mĪ» relation we had for a particle with zero rest mass. Of course, it does: the Ī» = h/mc relation is, once again, a limit for vg going to c. By the way, it is interesting to note that the vpĀ·vgĀ = c2 relation implies that the phase velocity is always superluminal. That’ easy to see when you re-write the equation in terms ofĀ relativeĀ velocities: (vp/c)Ā·(vg/c) =Ā Ī²phaseĀ·Ī²groupĀ = 1. Hence, ifĀ Ī²groupĀ < 1, thenĀ Ī²phaseĀ > 1.

So whatĀ isĀ the geometry,Ā really? Letās look at the Ļ = aĀ·cos(pāx/Ä§ – Eāt/Ä§) iĀ·aĀ·sin(pāx/Ä§ – Eāt/Ä§) formula once more. If we write pāx/Ä§ as Ī, then we will be interested to know for what x this phase factor will be equal to 2Ļ. So we write:

Ī =pāx/Ä§ = 2Ļ āĀ x = 2ĻāÄ§/p = h/p = Ī»Ā Ā

So now we get a meaningful interpretation for that wavelength: it’s that distance between the crests of the wave, so to speak, as illustrated below.

Can we now find a meaningful (i.e.Ā geometric) interpretation for the group and phase velocity? If you look at the illustration above, you see we can sort of distinguish (1) a linear velocity (the speed with which those wave crests move) and (2) some kind of circular or tangential velocity (the velocity along the red contour lineĀ above). Weāll probably need the formula for the tangential velocity: v = aāĻ. If p = 0 (so we have that weird infinitesimally long wavelength), then we have two velocities:

1. The tangential velocity around theĀ aĀ·eiĀ·EĀ·tĀ  circle, so to speak, and that will just be equal toĀ v = aāĻ =Ā aāE/Ä§.
2. The red contour line sort of gets stretched out, like infinitely long, and the velocity becomes… What does it do? Does it go toĀ ā , or toĀ c?

Let’s think about this. For a particle at rest, we had this weird calculation. We had an angular momentum formula (for an electron) which we equated with theĀ real-life +Ä§/2 or āÄ§/2 values of its spin. And so we got aĀ numericalĀ value forĀ a. It was the Compton radius: the scattering radius for an electron. Let me copy it once again:

Just to bring this story a bit back to Earth, you should note the calculated value:Ā aĀ = 3.8616Ć10ā13 m.Ā We did then another weird calculation. We said all of the energy of the electron had to be packed in thisĀ cylinderĀ that might of might not be there. The point was: the energy is finite, so thatĀ elementaryĀ wavefunction cannotĀ have an infinite length in space. Indeed, assuming that the energy was distributed uniformly, we jotted down this formula, which reflects the formula for theĀ volume of a cylinder:

E = ĻĀ·a2Ā·l āĀ lĀ = E/(ĻĀ·a2)

Using the value we got for the Compton scattering radius (aĀ =Ā 3.8616Ć10ā13 m), we got an astronomical value forĀ l. Let me write it out:

lĀ =Ā (8.19Ć10ā14)/(ĻĀ·14.9Ć10ā26) ā 0.175Ć1012Ā m

It is,Ā literally, an astronomical value:Ā 0.175Ć1012Ā m is 175 millionĀ kilometer, so that’s like theĀ distance between the Sun and the Earth. We wrote, jokingly, that such space is too large to look for an electron and, hence, that we should really build a proper packet by making use of the Uncertainty Principle: allowing for uncertainty in the energy should, effectively, reduce the uncertainty in position.

But… Well… What if we use that value as the value forĀ Ī»? We’d get that linear velocity, right? Let’s try it. TheĀ periodĀ is equal to T =Ā T = 2ĻĀ·(Ä§/E) = h/E and Ī» =Ā E/(ĻĀ·a2), so we write:We can write this as a function of m and theĀ cĀ andĀ Ä§ constants only:

A weird formula but not necessarily nonsensical: we get a finite speed. Now, if the wavelength becomes somewhat less astronomical, we’ll get different values of course. I have a strange feeling that, with these formula, we should, somehow, be able to explain relativistic length contraction. But I will let you think about that as for now. Here I just wanted toĀ showĀ the geometry of the wavefunction a bit more in detail.

[1]Ā The discussions on the mass of neutrinos are interesting in this regard. Scientists all felt the neutrinoĀ had toĀ haveĀ some (rest) mass, so my instinct on this is theirs. In fact, only recently experimental confirmation came in, and the mass of the known neutrino flavors was estimated to be something like 0.12 eV/c2. This mass combines the three known neutrino flavors. To understand this number, you should note it is the same order of magnitude of the equivalent mass of low-energy photons, like infrared or microwave radiation.

# Wavefunctions as gravitational waves

This is the paper I always wanted to write. It is there now, and I think it is good – and that‘s an understatement. š It is probably best to download it as a pdf-file from the viXra.org site because this was a rather fast ‘copy and paste’ job from the Word version of the paper, so there may be issues with boldface notation (vector notation), italics and, most importantly, with formulas – which I, sadly, have to ‘snip’ into this WordPress blog, as they don’t have an easy copy function for mathematical formulas.

It’s great stuff. If you have been following my blog – and many of you have – you will want to digest this. š

Abstract : This paper explores the implications of associating the components of the wavefunction with a physical dimension: force per unit mass ā which is, of course, the dimension of acceleration (m/s2) and gravitational fields. The classical electromagnetic field equations for energy densities, the Poynting vector and spin angular momentum are then re-derived by substituting the electromagnetic N/C unit of field strength (mass per unit charge) by the new N/kg = m/s2 dimension.

The results are elegant and insightful. For example, the energy densities are proportional to the square of the absolute value of the wavefunction and, hence, to the probabilities, which establishes a physical normalization condition. Also, SchrĆ¶dingerās wave equation may then, effectively, be interpreted as a diffusion equation for energy, and the wavefunction itself can be interpreted as a propagating gravitational wave. Finally, as an added bonus, concepts such as the Compton scattering radius for a particle, spin angular momentum, and the boson-fermion dichotomy, can also be explained more intuitively.

While the approach offers a physical interpretation of the wavefunction, the author argues that the core of the Copenhagen interpretations revolves around the complementarity principle, which remains unchallenged because the interpretation of amplitude waves as traveling fields does not explain the particle nature of matter.

# Introduction

This is not another introduction to quantum mechanics. We assume the reader is already familiar with the key principles and, importantly, with the basic math. We offer an interpretation of wave mechanics. As such, we do not challenge the complementarity principle: the physical interpretation of the wavefunction that is offered here explains the wave nature of matter only. It explains diffraction and interference of amplitudes but it does not explain why a particle will hit the detector not as a wave but as a particle. Hence, the Copenhagen interpretation of the wavefunction remains relevant: we just push its boundaries.

The basic ideas in this paper stem from a simple observation: the geometric similarity between the quantum-mechanical wavefunctions and electromagnetic waves is remarkably similar. The components of both waves are orthogonal to the direction of propagation and to each other. Only the relative phase differs : the electric and magnetic field vectors (E and B) have the same phase. In contrast, the phase of the real and imaginary part of the (elementary) wavefunction (Ļ = aĀ·eāiāĪø = aācosĪø – aāsinĪø) differ by 90 degrees (Ļ/2).[1] Pursuing the analogy, we explore the following question: if the oscillating electric and magnetic field vectors of an electromagnetic wave carry the energy that one associates with the wave, can we analyze the real and imaginary part of the wavefunction in a similar way?

We show the answer is positive and remarkably straightforward.  If the physical dimension of the electromagnetic field is expressed in newton per coulomb (force per unit charge), then the physical dimension of the components of the wavefunction may be associated with force per unit mass (newton per kg).[2] Of course, force over some distance is energy. The question then becomes: what is the energy concept here? Kinetic? Potential? Both?

The similarity between the energy of a (one-dimensional) linear oscillator (E = mĀ·a2Ā·Ļ2/2) and Einsteinās relativistic energy equation E = māc2 inspires us to interpret the energy as a two-dimensional oscillation of mass. To assist the reader, we construct a two-piston engine metaphor.[3] We then adapt the formula for the electromagnetic energy density to calculate the energy densities for the wave function. The results are elegant and intuitive: the energy densities are proportional to the square of the absolute value of the wavefunction and, hence, to the probabilities. SchrĆ¶dingerās wave equation may then, effectively, be interpreted as a diffusion equation for energy itself.

As an added bonus, concepts such as the Compton scattering radius for a particle and spin angular, as well as the boson-fermion dichotomy can be explained in a fully intuitive way.[4]

Of course, such interpretation is also an interpretation of the wavefunction itself, and the immediate reaction of the reader is predictable: the electric and magnetic field vectors are, somehow, to be looked at as real vectors. In contrast, the real and imaginary components of the wavefunction are not. However, this objection needs to be phrased more carefully. First, it may be noted that, in a classical analysis, the magnetic force is a pseudovector itself.[5] Second, a suitable choice of coordinates may make quantum-mechanical rotation matrices irrelevant.[6]

Therefore, the author is of the opinion that this little paper may provide some fresh perspective on the question, thereby further exploring Einsteinās basic sentiment in regard to quantum mechanics, which may be summarized as follows: there must be some physical explanation for the calculated probabilities.[7]

We will, therefore, start with Einsteinās relativistic energy equation (E = mc2) and wonder what it could possibly tell us.

# I. Energy as a two-dimensional oscillation of mass

The structural similarity between the relativistic energy formula, the formula for the total energy of an oscillator, and the kinetic energy of a moving body, is striking:

1. E = mc2
2. E = mĻ2/2
3. E = mv2/2

In these formulas, Ļ, v and c all describe some velocity.[8] Of course, there is the 1/2 factor in the E = mĻ2/2 formula[9], but that is exactly the point we are going to explore here: can we think of an oscillation in two dimensions, so it stores an amount of energy that is equal to E = 2Ā·mĀ·Ļ2/2 = mĀ·Ļ2?

That is easy enough. Think, for example, of a V-2 engine with the pistons at a 90-degree angle, as illustrated below. The 90Ā° angle makes it possible to perfectly balance the counterweight and the pistons, thereby ensuring smooth travel at all times. With permanently closed valves, the air inside the cylinder compresses and decompresses as the pistons move up and down and provides, therefore, a restoring force. As such, it will store potential energy, just like a spring, and the motion of the pistons will also reflect that of a mass on a spring. Hence, we can describe it by a sinusoidal function, with the zero point at the center of each cylinder. We can, therefore, think of the moving pistons as harmonic oscillators, just like mechanical springs.

Figure 1: Oscillations in two dimensions

If we assume there is no friction, we have a perpetuum mobile here. The compressed air and the rotating counterweight (which, combined with the crankshaft, acts as a flywheel[10]) store the potential energy. The moving masses of the pistons store the kinetic energy of the system.[11]

At this point, it is probably good to quickly review the relevant math. If the magnitude of the oscillation is equal to a, then the motion of the piston (or the mass on a spring) will be described by x = aĀ·cos(ĻĀ·t + Ī).[12] Needless to say, Ī is just a phase factor which defines our t = 0 point, and Ļ is the natural angular frequency of our oscillator. Because of the 90Ā° angle between the two cylinders, Ī would be 0 for one oscillator, and āĻ/2 for the other. Hence, the motion of one piston is given by x = aĀ·cos(ĻĀ·t), while the motion of the other is given by x = aĀ·cos(ĻĀ·tāĻ/2) = aĀ·sin(ĻĀ·t).

The kinetic and potential energy of one oscillator (think of one piston or one spring only) can then be calculated as:

1. K.E. = T = mĀ·v2/2 = (1/2)Ā·mĀ·Ļ2Ā·a2Ā·sin2(ĻĀ·t + Ī)
2. P.E. = U = kĀ·x2/2 = (1/2)Ā·kĀ·a2Ā·cos2(ĻĀ·t + Ī)

The coefficient k in the potential energy formula characterizes the restoring force: F = ākĀ·x. From the dynamics involved, it is obvious that k must be equal to mĀ·Ļ2. Hence, the total energy is equal to:

E = T + U = (1/2)Ā· mĀ·Ļ2Ā·a2Ā·[sin2(ĻĀ·t + Ī) + cos2(ĻĀ·t + Ī)] = mĀ·a2Ā·Ļ2/2

To facilitate the calculations, we will briefly assume k = mĀ·Ļ2 and a are equal to 1. The motion of our first oscillator is given by the cos(ĻĀ·t) = cosĪø function (Īø = ĻĀ·t), and its kinetic energy will be equal to sin2Īø. Hence, the (instantaneous) change in kinetic energy at any point in time will be equal to:

d(sin2Īø)/dĪø = 2āsinĪøād(sinĪø)/dĪø = 2āsinĪøācosĪø

Let us look at the second oscillator now. Just think of the second piston going up and down in the V-2 engine. Its motion is given by the sinĪø function, which is equal to cos(ĪøāĻ /2). Hence, its kinetic energy is equal to sin2(ĪøāĻ /2), and how it changes ā as a function of Īø ā will be equal to:

2āsin(ĪøāĻ /2)ācos(ĪøāĻ /2) = = ā2ācosĪøāsinĪø = ā2āsinĪøācosĪø

We have our perpetuum mobile! While transferring kinetic energy from one piston to the other, the crankshaft will rotate with a constant angular velocity: linear motion becomes circular motion, and vice versa, and the total energy that is stored in the system is T + U = ma2Ļ2.

We have a great metaphor here. Somehow, in this beautiful interplay between linear and circular motion, energy is borrowed from one place and then returns to the other, cycle after cycle. We know the wavefunction consist of a sine and a cosine: the cosine is the real component, and the sine is the imaginary component. Could they be equally real? Could each represent half of the total energy of our particle? Should we think of the c in our E = mc2 formula as an angular velocity?

These are sensible questions. Let us explore them.

# II. The wavefunction as a two-dimensional oscillation

The elementary wavefunction is written as:

Ļ = aĀ·eāi[EĀ·t ā pāx]/Ä§aĀ·eāi[EĀ·t ā pāx]/Ä§ = aĀ·cos(pāx/Ä§ Eāt/Ä§) + iĀ·aĀ·sin(pāx/Ä§ Eāt/Ä§)

When considering a particle at rest (p = 0) this reduces to:

Ļ = aĀ·eāiāEĀ·t/Ä§ = aĀ·cos(Eāt/Ä§) + iĀ·aĀ·sin(Eāt/Ä§) = aĀ·cos(Eāt/Ä§) iĀ·aĀ·sin(Eāt/Ä§)

Let us remind ourselves of the geometry involved, which is illustrated below. Note that the argument of the wavefunction rotates clockwise with time, while the mathematical convention for measuring the phase angle (Ļ) is counter-clockwise.

Figure 2: Eulerās formula

If we assume the momentum p is all in the x-direction, then the p and x vectors will have the same direction, and pāx/Ä§ reduces to pāx/Ä§. Most illustrations ā such as the one below ā will either freeze x or, else, t. Alternatively, one can google web animations varying both. The point is: we also have a two-dimensional oscillation here. These two dimensions are perpendicular to the direction of propagation of the wavefunction. For example, if the wavefunction propagates in the x-direction, then the oscillations are along the y– and z-axis, which we may refer to as the real and imaginary axis. Note how the phase difference between the cosine and the sine  ā the real and imaginary part of our wavefunction ā appear to give some spin to the whole. I will come back to this.

Figure 3: Geometric representation of the wavefunction

Hence, if we would say these oscillations carry half of the total energy of the particle, then we may refer to the real and imaginary energy of the particle respectively, and the interplay between the real and the imaginary part of the wavefunction may then describe how energy propagates through space over time.

Let us consider, once again, a particle at rest. Hence, p = 0 and the (elementary) wavefunction reduces to Ļ = aĀ·eāiāEĀ·t/Ä§. Hence, the angular velocity of both oscillations, at some point x, is given by Ļ = -E/Ä§. Now, the energy of our particle includes all of the energy ā kinetic, potential and rest energy ā and is, therefore, equal to E = mc2.

Can we, somehow, relate this to the mĀ·a2Ā·Ļ2 energy formula for our V-2 perpetuum mobile? Our wavefunction has an amplitude too. Now, if the oscillations of the real and imaginary wavefunction store the energy of our particle, then their amplitude will surely matter. In fact, the energy of an oscillation is, in general, proportional to the square of the amplitude: E Āµ a2. We may, therefore, think that the a2 factor in the E = mĀ·a2Ā·Ļ2 energy will surely be relevant as well.

However, here is a complication: an actual particle is localized in space and can, therefore, not be represented by the elementary wavefunction. We must build a wave packet for that: a sum of wavefunctions, each with their own amplitude ak, and their own Ļi = -Ei/Ä§. Each of these wavefunctions will contribute some energy to the total energy of the wave packet. To calculate the contribution of each wave to the total, both ai as well as Ei will matter.

What is Ei? Ei varies around some average E, which we can associate with some average mass m: m = E/c2. The Uncertainty Principle kicks in here. The analysis becomes more complicated, but a formula such as the one below might make sense:We can re-write this as:What is the meaning of this equation? We may look at it as some sort of physical normalization condition when building up the Fourier sum. Of course, we should relate this to the mathematical normalization condition for the wavefunction. Our intuition tells us that the probabilities must be related to the energy densities, but how exactly? We will come back to this question in a moment. Let us first think some more about the enigma: what is mass?

Before we do so, let us quickly calculate the value of c2Ä§2: it is about 1Ā“1051 N2ām4. Let us also do a dimensional analysis: the physical dimensions of the E = mĀ·a2Ā·Ļ2 equation make sense if we express m in kg, a in m, and Ļ in rad/s. We then get: [E] = kgām2/s2 = (Nās2/m)ām2/s2 = Nām = J. The dimensions of the left- and right-hand side of the physical normalization condition is N3ām5.

# III. What is mass?

We came up, playfully, with a meaningful interpretation for energy: it is a two-dimensional oscillation of mass. But what is mass? A new aether theory is, of course, not an option, but then what is it that is oscillating? To understand the physics behind equations, it is always good to do an analysis of the physical dimensions in the equation. Let us start with Einsteinās energy equation once again. If we want to look at mass, we should re-write it as m = E/c2:

[m] = [E/c2] = J/(m/s)2 = NĀ·mās2/m2 = NĀ·s2/m = kg

This is not very helpful. It only reminds us of Newtonās definition of a mass: mass is that what gets accelerated by a force. At this point, we may want to think of the physical significance of the absolute nature of the speed of light. Einsteinās E = mc2 equation implies we can write the ratio between the energy and the mass of any particle is always the same, so we can write, for example:This reminds us of the Ļ2= C1/L or Ļ2 = k/m of harmonic oscillators once again.[13] The key difference is that the Ļ2= C1/L and Ļ2 = k/m formulas introduce two or more degrees of freedom.[14] In contrast, c2= E/m for any particle, always. However, that is exactly the point: we can modulate the resistance, inductance and capacitance of electric circuits, and the stiffness of springs and the masses we put on them, but we live in one physical space only: our spacetime. Hence, the speed of light c emerges here as the defining property of spacetime ā the resonant frequency, so to speak. We have no further degrees of freedom here.

The Planck-Einstein relation (for photons) and the de Broglie equation (for matter-particles) have an interesting feature: both imply that the energy of the oscillation is proportional to the frequency, with Planckās constant as the constant of proportionality. Now, for one-dimensional oscillations ā think of a guitar string, for example ā we know the energy will be proportional to the square of the frequency. It is a remarkable observation: the two-dimensional matter-wave, or the electromagnetic wave, gives us two waves for the price of one, so to speak, each carrying half of the total energy of the oscillation but, as a result, we get a proportionality between E and f instead of between E and f2.

However, such reflections do not answer the fundamental question we started out with: what is mass? At this point, it is hard to go beyond the circular definition that is implied by Einsteinās formula: energy is a two-dimensional oscillation of mass, and mass packs energy, and c emerges us as the property of spacetime that defines how exactly.

When everything is said and done, this does not go beyond stating that mass is some scalar field. Now, a scalar field is, quite simply, some real number that we associate with a position in spacetime. The Higgs field is a scalar field but, of course, the theory behind it goes much beyond stating that we should think of mass as some scalar field. The fundamental question is: why and how does energy, or matter, condense into elementary particles? That is what the Higgs mechanism is about but, as this paper is exploratory only, we cannot even start explaining the basics of it.

What we can do, however, is look at the wave equation again (SchrĆ¶dingerās equation), as we can now analyze it as an energy diffusion equation.

# IV. SchrĆ¶dingerās equation as an energy diffusion equation

The interpretation of SchrĆ¶dingerās equation as a diffusion equation is straightforward. Feynman (Lectures, III-16-1) briefly summarizes it as follows:

āWe can think of SchrĆ¶dingerās equation as describing the diffusion of the probability amplitude from one point to the next. [ā¦] But the imaginary coefficient in front of the derivative makes the behavior completely different from the ordinary diffusion such as you would have for a gas spreading out along a thin tube. Ordinary diffusion gives rise to real exponential solutions, whereas the solutions of SchrĆ¶dingerās equation are complex waves.ā[17]

Let us review the basic math. For a particle moving in free space ā with no external force fields acting on it ā there is no potential (U = 0) and, therefore, the UĻ term disappears. Therefore, SchrĆ¶dingerās equation reduces to:

āĻ(x, t)/āt = iĀ·(1/2)Ā·(Ä§/meff)Ā·ā2Ļ(x, t)

The ubiquitous diffusion equation in physics is:

āĻ(x, t)/āt = DĀ·ā2Ļ(x, t)

The structural similarity is obvious. The key difference between both equations is that the wave equation gives us two equations for the price of one. Indeed, because Ļ is a complex-valued function, with a real and an imaginary part, we get the following equations[18]:

1. Re(āĻ/āt) = ā(1/2)Ā·(Ä§/meff)Ā·Im(ā2Ļ)
2. Im(āĻ/āt) = (1/2)Ā·(Ä§/meff)Ā·Re(ā2Ļ)

These equations make us think of the equations for an electromagnetic wave in free space (no stationary charges or currents):

1. āB/āt = āāĆE
2. āE/āt = c2āĆB

The above equations effectively describe a propagation mechanism in spacetime, as illustrated below.

Figure 4: Propagation mechanisms

The Laplacian operator (ā2), when operating on a scalar quantity, gives us a flux density, i.e. something expressed per square meter (1/m2). In this case, it is operating on Ļ(x, t), so what is the dimension of our wavefunction Ļ(x, t)? To answer that question, we should analyze the diffusion constant in SchrĆ¶dingerās equation, i.e. the (1/2)Ā·(Ä§/meff) factor:

1. As a mathematical constant of proportionality, it will quantify the relationship between both derivatives (i.e. the time derivative and the Laplacian);
2. As a physical constant, it will ensure the physical dimensions on both sides of the equation are compatible.

Now, the Ä§/meff factor is expressed in (NĀ·mĀ·s)/(NĀ· s2/m) = m2/s. Hence, it does ensure the dimensions on both sides of the equation are, effectively, the same: āĻ/āt is a time derivative and, therefore, its dimension is s1 while, as mentioned above, the dimension of ā2Ļ is m2. However, this does not solve our basic question: what is the dimension of the real and imaginary part of our wavefunction?

At this point, mainstream physicists will say: it does not have a physical dimension, and there is no geometric interpretation of SchrĆ¶dingerās equation. One may argue, effectively, that its argument, (pāx – Eāt)/Ä§, is just a number and, therefore, that the real and imaginary part of Ļ is also just some number.

To this, we may object that Ä§ may be looked as a mathematical scaling constant only. If we do that, then the argument of Ļ will, effectively, be expressed in action units, i.e. in NĀ·mĀ·s. It then does make sense to also associate a physical dimension with the real and imaginary part of Ļ. What could it be?

We may have a closer look at Maxwellās equations for inspiration here. The electric field vector is expressed in newton (the unit of force) per unit of charge (coulomb). Now, there is something interesting here. The physical dimension of the magnetic field is N/C divided by m/s.[19] We may write B as the following vector cross-product: B = (1/c)āexĆE, with ex the unit vector pointing in the x-direction (i.e. the direction of propagation of the wave). Hence, we may associate the (1/c)āexĆ operator, which amounts to a rotation by 90 degrees, with the s/m dimension. Now, multiplication by i also amounts to a rotation by 90Ā° degrees. Hence, we may boldly write: B = (1/c)āexĆE = (1/c)āiāE. This allows us to also geometrically interpret SchrĆ¶dingerās equation in the way we interpreted it above (see Figure 3).[20]

Still, we have not answered the question as to what the physical dimension of the real and imaginary part of our wavefunction should be. At this point, we may be inspired by the structural similarity between Newtonās and Coulombās force laws:Hence, if the electric field vector E is expressed in force per unit charge (N/C), then we may want to think of associating the real part of our wavefunction with a force per unit mass (N/kg). We can, of course, do a substitution here, because the mass unit (1 kg) is equivalent to 1 NĀ·s2/m. Hence, our N/kg dimension becomes:

N/kg = N/(NĀ·s2/m)= m/s2

What is this: m/s2? Is that the dimension of the aĀ·cosĪø term in the aĀ·eāiĪø aĀ·cosĪø ā iĀ·aĀ·sinĪø wavefunction?

My answer is: why not? Think of it: m/s2 is the physical dimension of acceleration: the increase or decrease in velocity (m/s) per second. It ensures the wavefunction for any particle ā matter-particles or particles with zero rest mass (photons) ā and the associated wave equation (which has to be the same for all, as the spacetime we live in is one) are mutually consistent.

In this regard, we should think of how we would model a gravitational wave. The physical dimension would surely be the same: force per mass unit. It all makes sense: wavefunctions may, perhaps, be interpreted as traveling distortions of spacetime, i.e. as tiny gravitational waves.

# V. Energy densities and flows

Pursuing the geometric equivalence between the equations for an electromagnetic wave and SchrĆ¶dingerās equation, we can now, perhaps, see if there is an equivalent for the energy density. For an electromagnetic wave, we know that the energy density is given by the following formula:E and B are the electric and magnetic field vector respectively. The Poynting vector will give us the directional energy flux, i.e. the energy flow per unit area per unit time. We write:Needless to say, the āā operator is the divergence and, therefore, gives us the magnitude of a (vector) fieldās source or sink at a given point. To be precise, the divergence gives us the volume density of the outward flux of a vector field from an infinitesimal volume around a given point. In this case, it gives us the volume density of the flux of S.

We can analyze the dimensions of the equation for the energy density as follows:

1. E is measured in newton per coulomb, so [EāE] = [E2] = N2/C2.
2. B is measured in (N/C)/(m/s), so we get [BāB] = [B2] = (N2/C2)Ā·(s2/m2). However, the dimension of our c2 factor is (m2/s2) and so weāre also left with N2/C2.
3. The Ļµ0 is the electric constant, aka as the vacuum permittivity. As a physical constant, it should ensure the dimensions on both sides of the equation work out, and they do: [Īµ0] = C2/(NĀ·m2) and, therefore, if we multiply that with N2/C2, we find that is expressed in J/m3.[21]

Replacing the newton per coulomb unit (N/C) by the newton per kg unit (N/kg) in the formulas above should give us the equivalent of the energy density for the wavefunction. We just need to substitute Ļµ0 for an equivalent constant. We may to give it a try. If the energy densities can be calculated ā which are also mass densities, obviously ā then the probabilities should be proportional to them.

Let us first see what we get for a photon, assuming the electromagnetic wave represents its wavefunction. Substituting B for (1/c)āiāE or for ā(1/c)āiāE gives us the following result:Zero!? An unexpected result! Or not? We have no stationary charges and no currents: only an electromagnetic wave in free space. Hence, the local energy conservation principle needs to be respected at all points in space and in time. The geometry makes sense of the result: for an electromagnetic wave, the magnitudes of E and B reach their maximum, minimum and zero point simultaneously, as shown below.[22] This is because their phase is the same.

Figure 5: Electromagnetic wave: E and B

Should we expect a similar result for the energy densities that we would associate with the real and imaginary part of the matter-wave? For the matter-wave, we have a phase difference between aĀ·cosĪø and aĀ·sinĪø, which gives a different picture of the propagation of the wave (see Figure 3).[23] In fact, the geometry of the suggestion suggests some inherent spin, which is interesting. I will come back to this. Let us first guess those densities. Making abstraction of any scaling constants, we may write:We get what we hoped to get: the absolute square of our amplitude is, effectively, an energy density !

|Ļ|2  = |aĀ·eāiāEĀ·t/Ä§|2 = a2 = u

This is very deep. A photon has no rest mass, so it borrows and returns energy from empty space as it travels through it. In contrast, a matter-wave carries energy and, therefore, has some (rest) mass. It is therefore associated with an energy density, and this energy density gives us the probabilities. Of course, we need to fine-tune the analysis to account for the fact that we have a wave packet rather than a single wave, but that should be feasible.

As mentioned, the phase difference between the real and imaginary part of our wavefunction (a cosine and a sine function) appear to give some spin to our particle. We do not have this particularity for a photon. Of course, photons are bosons, i.e. spin-zero particles, while elementary matter-particles are fermions with spin-1/2. Hence, our geometric interpretation of the wavefunction suggests that, after all, there may be some more intuitive explanation of the fundamental dichotomy between bosons and fermions, which puzzled even Feynman:

āWhy is it that particles with half-integral spin are Fermi particles, whereas particles with integral spin are Bose particles? We apologize for the fact that we cannot give you an elementary explanation. An explanation has been worked out by Pauli from complicated arguments of quantum field theory and relativity. He has shown that the two must necessarily go together, but we have not been able to find a way of reproducing his arguments on an elementary level. It appears to be one of the few places in physics where there is a rule which can be stated very simply, but for which no one has found a simple and easy explanation. The explanation is deep down in relativistic quantum mechanics. This probably means that we do not have a complete understanding of the fundamental principle involved.ā (Feynman, Lectures, III-4-1)

The physical interpretation of the wavefunction, as presented here, may provide some better understanding of āthe fundamental principle involvedā: the physical dimension of the oscillation is just very different. That is all: it is force per unit charge for photons, and force per unit mass for matter-particles. We will examine the question of spin somewhat more carefully in section VII. Let us first examine the matter-wave some more.

# VI. Group and phase velocity of the matter-wave

The geometric representation of the matter-wave (see Figure 3) suggests a traveling wave and, yes, of course: the matter-wave effectively travels through space and time. But what is traveling, exactly? It is the pulse ā or the signal ā only: the phase velocity of the wave is just a mathematical concept and, even in our physical interpretation of the wavefunction, the same is true for the group velocity of our wave packet. The oscillation is two-dimensional, but perpendicular to the direction of travel of the wave. Hence, nothing actually moves with our particle.

Here, we should also reiterate that we did not answer the question as to what is oscillating up and down and/or sideways: we only associated a physical dimension with the components of the wavefunction ā newton per kg (force per unit mass), to be precise. We were inspired to do so because of the physical dimension of the electric and magnetic field vectors (newton per coulomb, i.e. force per unit charge) we associate with electromagnetic waves which, for all practical purposes, we currently treat as the wavefunction for a photon. This made it possible to calculate the associated energy densities and a Poynting vector for energy dissipation. In addition, we showed that SchrĆ¶dinger’s equation itself then becomes a diffusion equation for energy. However, let us now focus some more on the asymmetry which is introduced by the phase difference between the real and the imaginary part of the wavefunction. Look at the mathematical shape of the elementary wavefunction once again:

Ļ = aĀ·eāi[EĀ·t ā pāx]/Ä§aĀ·eāi[EĀ·t ā pāx]/Ä§ = aĀ·cos(pāx/Ä§ ā Eāt/Ä§) + iĀ·aĀ·sin(pāx/Ä§ ā Eāt/Ä§)

The minus sign in the argument of our sine and cosine function defines the direction of travel: an F(xāvāt) wavefunction will always describe some wave that is traveling in the positive x-direction (with the wave velocity), while an F(x+vāt) wavefunction will travel in the negative x-direction. For a geometric interpretation of the wavefunction in three dimensions, we need to agree on how to define i or, what amounts to the same, a convention on how to define clockwise and counterclockwise directions: if we look at a clock from the back, then its hand will be moving counterclockwise. So we need to establish the equivalent of the right-hand rule. However, let us not worry about that now. Let us focus on the interpretation. To ease the analysis, we’ll assume we’re looking at a particle at rest. Hence, p = 0, and the wavefunction reduces to:

Ļ = aĀ·eāiāEĀ·t/Ä§ = aĀ·cos(āEāt/Ä§) + iĀ·aĀ·sin(āE0āt/Ä§) = aĀ·cos(E0āt/Ä§) ā iĀ·aĀ·sin(E0āt/Ä§)

E0 is, of course, the rest mass of our particle and, now that we are here, we should probably wonder whose time we are talking about: is it our time, or is the proper time of our particle? Well… In this situation, we are both at rest so it does not matter: t is, effectively, the proper time so perhaps we should write it as t0. It does not matter. You can see what we expect to see: E0/Ä§ pops up as the natural frequency of our matter-particle: (E0/Ä§)āt = Ļāt. Remembering the Ļ = 2ĻĀ·f = 2Ļ/T and T = 1/formulas, we can associate a period and a frequency with this wave, using the Ļ = 2ĻĀ·f = 2Ļ/T. Noting that Ä§ = h/2Ļ, we find the following:

T = 2ĻĀ·(Ä§/E0) = h/E0 ā = E0/h = m0c2/h

This is interesting, because we can look at the period as a natural unit of time for our particle. What about the wavelength? That is tricky because we need to distinguish between group and phase velocity here. The group velocity (vg) should be zero here, because we assume our particle does not move. In contrast, the phase velocity is given by vp = Ī»Ā·= (2Ļ/k)Ā·(Ļ/2Ļ) = Ļ/k. In fact, we’ve got something funny here: the wavenumber k = p/Ä§ is zero, because we assume the particle is at rest, so p = 0. So we have a division by zero here, which is rather strange. What do we get assuming the particle is not at rest? We write:

vp = Ļ/k = (E/Ä§)/(p/Ä§) = E/p = E/(mĀ·vg) = (mĀ·c2)/(mĀ·vg) = c2/vg

This is interesting: it establishes a reciprocal relation between the phase and the group velocity, with as a simple scaling constant. Indeed, the graph below shows the shape of the function does not change with the value of c, and we may also re-write the relation above as:

vp/= Ī²p = c/vp = 1/Ī²g = 1/(c/vp)

Figure 6: Reciprocal relation between phase and group velocity

We can also write the mentioned relationship as vpĀ·vg = c2, which reminds us of the relationship between the electric and magnetic constant (1/Īµ0)Ā·(1/Ī¼0) = c2. This is interesting in light of the fact we can re-write this as (cĀ·Īµ0)Ā·(cĀ·Ī¼0) = 1, which shows electricity and magnetism are just two sides of the same coin, so to speak.[24]

Interesting, but how do we interpret the math? What about the implications of the zero value for wavenumber k = p/Ä§. We would probably like to think it implies the elementary wavefunction should always be associated with some momentum, because the concept of zero momentum clearly leads to weird math: something times zero cannot be equal to c2! Such interpretation is also consistent with the Uncertainty Principle: if ĪxĀ·Īp ā„ Ä§, then neither Īx nor Īp can be zero. In other words, the Uncertainty Principle tells us that the idea of a pointlike particle actually being at some specific point in time and in space does not make sense: it has to move. It tells us that our concept of dimensionless points in time and space are mathematical notions only. Actual particles – including photons – are always a bit spread out, so to speak, and – importantly – they have to move.

For a photon, this is self-evident. It has no rest mass, no rest energy, and, therefore, it is going to move at the speed of light itself. We write: p = mĀ·c = mĀ·c2/= E/c. Using the relationship above, we get:

vp = Ļ/k = (E/Ä§)/(p/Ä§) = E/p = c ā vg = c2/vp = c2/c = c

This is good: we started out with some reflections on the matter-wave, but here we get an interpretation of the electromagnetic wave as a wavefunction for the photon. But let us get back to our matter-wave. In regard to our interpretation of a particle having to move, we should remind ourselves, once again, of the fact that an actual particle is always localized in space and that it can, therefore, not be represented by the elementary wavefunction Ļ = aĀ·eāi[EĀ·t ā pāx]/Ä§ or, for a particle at rest, the Ļ = aĀ·eāiāEĀ·t/Ä§ function. We must build a wave packet for that: a sum of wavefunctions, each with their own amplitude ai, and their own Ļi = āEi/Ä§. Indeed, in section II, we showed that each of these wavefunctions will contribute some energy to the total energy of the wave packet and that, to calculate the contribution of each wave to the total, both ai as well as Ei matter. This may or may not resolve the apparent paradox. Let us look at the group velocity.

To calculate a meaningful group velocity, we must assume the vg = āĻi/āki = ā(Ei/Ä§)/ā(pi/Ä§) = ā(Ei)/ā(pi) exists. So we must have some dispersion relation. How do we calculate it? We need to calculate Ļi as a function of ki here, or Ei as a function of pi. How do we do that? Well… There are a few ways to go about it but one interesting way of doing it is to re-write SchrĆ¶dinger’s equation as we did, i.e. by distinguishing the real and imaginary parts of the āĻ/āt =iĀ·[Ä§/(2m)]Ā·ā2Ļ wave equation and, hence, re-write it as the following pair of two equations:

1. Re(āĻ/āt) = ā[Ä§/(2meff)]Ā·Im(ā2Ļ) ā ĻĀ·cos(kx ā Ļt) = k2Ā·[Ä§/(2meff)]Ā·cos(kx ā Ļt)
2. Im(āĻ/āt) = [Ä§/(2meff)]Ā·Re(ā2Ļ) ā ĻĀ·sin(kx ā Ļt) = k2Ā·[Ä§/(2meff)]Ā·sin(kx ā Ļt)

Both equations imply the following dispersion relation:

Ļ = Ä§Ā·k2/(2meff)

Of course, we need to think about the subscripts now: we have Ļi, ki, but… What about meff or, dropping the subscript, m? Do we write it as mi? If so, what is it? Well… It is the equivalent mass of Ei obviously, and so we get it from the mass-energy equivalence relation: mi = Ei/c2. It is a fine point, but one most people forget about: they usually just write m. However, if there is uncertainty in the energy, then Einstein’s mass-energy relation tells us we must have some uncertainty in the (equivalent) mass too. Here, I should refer back to Section II: Ei varies around some average energy E and, therefore, the Uncertainty Principle kicks in.

# VII. Explaining spin

The elementary wavefunction vector ā i.e. the vector sum of the real and imaginary component ā rotates around the x-axis, which gives us the direction of propagation of the wave (see Figure 3). Its magnitude remains constant. In contrast, the magnitude of the electromagnetic vector ā defined as the vector sum of the electric and magnetic field vectors ā oscillates between zero and some maximum (see Figure 5).

We already mentioned that the rotation of the wavefunction vector appears to give some spin to the particle. Of course, a circularly polarized wave would also appear to have spin (think of the E and B vectors rotating around the direction of propagation – as opposed to oscillating up and down or sideways only). In fact, a circularly polarized light does carry angular momentum, as the equivalent mass of its energy may be thought of as rotating as well. But so here we are looking at a matter-wave.

The basic idea is the following: if we look at Ļ = aĀ·eāiāEĀ·t/Ä§ as some real vector ā as a two-dimensional oscillation of mass, to be precise ā then we may associate its rotation around the direction of propagation with some torque. The illustration below reminds of the math here.

Figure 7: Torque and angular momentum vectors

A torque on some mass about a fixed axis gives it angular momentum, which we can write as the vector cross-product L = rĆp or, perhaps easier for our purposes here as the product of an angular velocity (Ļ) and rotational inertia (I), aka as the moment of inertia or the angular mass. We write:

L = IĀ·Ļ

Note we can write L and Ļ in boldface here because they are (axial) vectors. If we consider their magnitudes only, we write L = IĀ·Ļ (no boldface). We can now do some calculations. Let us start with the angular velocity. In our previous posts, we showed that the period of the matter-wave is equal to T = 2ĻĀ·(Ä§/E0). Hence, the angular velocity must be equal to:

Ļ = 2Ļ/[2ĻĀ·(Ä§/E0)] = E0/Ä§

We also know the distance r, so that is the magnitude of r in the LrĆp vector cross-product: it is just a, so that is the magnitude of Ļ = aĀ·eāiāEĀ·t/Ä§. Now, the momentum (p) is the product of a linear velocity (v) – in this case, the tangential velocity – and some mass (m): p = mĀ·v. If we switch to scalar instead of vector quantities, then the (tangential) velocity is given by v = rĀ·Ļ. So now we only need to think about what we should use for m or, if we want to work with the angular velocity (Ļ), the angular mass (I). Here we need to make some assumption about the mass (or energy) distribution. Now, it may or may not sense to assume the energy in the oscillation ā and, therefore, the mass ā is distributed uniformly. In that case, we may use the formula for the angular mass of a solid cylinder: I = mĀ·r2/2. If we keep the analysis non-relativistic, then m = m0. Of course, the energy-mass equivalence tells us that m0 = E0/c2. Hence, this is what we get:

L = IĀ·Ļ = (m0Ā·r2/2)Ā·(E0/Ä§) = (1/2)Ā·a2Ā·(E0/c2)Ā·(E0/Ä§) = a2Ā·E02/(2Ā·Ä§Ā·c2)

Does it make sense? Maybe. Maybe not. Let us do a dimensional analysis: that wonāt check our logic, but it makes sure we made no mistakes when mapping mathematical and physical spaces. We have m2Ā·J2 = m2Ā·N2Ā·m2 in the numerator and NĀ·mĀ·sĀ·m2/s2 in the denominator. Hence, the dimensions work out: we get NĀ·mĀ·s as the dimension for L, which is, effectively, the physical dimension of angular momentum. It is also the action dimension, of course, and that cannot be a coincidence. Also note that the E = mc2 equation allows us to re-write it as:

L = a2Ā·E02/(2Ā·Ä§Ā·c2)

Of course, in quantum mechanics, we associate spin with the magnetic moment of a charged particle, not with its mass as such. Is there way to link the formula above to the one we have for the quantum-mechanical angular momentum, which is also measured in NĀ·mĀ·s units, and which can only take on one of two possible values: J = +Ä§/2 and āÄ§/2? It looks like a long shot, right? How do we go from (1/2)Ā·a2Ā·m02/Ä§ to Ā± (1/2)āÄ§? Let us do a numerical example. The energy of an electron is typically 0.510 MeV Ā» 8.1871Ć10ā14 Nām, and aā¦ What value should we take for a?

We have an obvious trio of candidates here: the Bohr radius, the classical electron radius (aka the Thompon scattering length), and the Compton scattering radius.

Let us start with the Bohr radius, so that is about 0.Ć10ā10 Nām. We get L = a2Ā·E02/(2Ā·Ä§Ā·c2) = 9.9Ć10ā31 Nāmās. Now that is about 1.88Ć104 times Ä§/2. That is a huge factor. The Bohr radius cannot be right: we are not looking at an electron in an orbital here. To show it does not make sense, we may want to double-check the analysis by doing the calculation in another way. We said each oscillation will always pack 6.626070040(81)Ć10ā34 joule in energy. So our electron should pack about 1.24Ć10ā20 oscillations. The angular momentum (L) we get when using the Bohr radius for a and the value of 6.626Ć10ā34 joule for E0 and the Bohr radius is equal to 6.49Ć10ā59 Nāmās. So that is the angular momentum per oscillation. When we multiply this with the number of oscillations (1.24Ć10ā20), we get about 8.01Ć10ā51 Nāmās, so that is a totally different number.

The classical electron radius is about 2.818Ć10ā15 m. We get an L that is equal to about 2.81Ć10ā39 Nāmās, so now it is a tiny fraction of Ä§/2! Hence, this leads us nowhere. Let us go for our last chance to get a meaningful result! Let us use the Compton scattering length, so that is about 2.42631Ć10ā12 m.

This gives us an L of 2.08Ć10ā33 Nāmās, which is only 20 times Ä§. This is not so bad, but it is good enough? Let us calculate it the other way around: what value should we take for a so as to ensure L = a2Ā·E02/(2Ā·Ä§Ā·c2) = Ä§/2? Let us write it out:

In fact, this is the formula for the so-called reduced Compton wavelength. This is perfect. We found what we wanted to find. Substituting this value for a (you can calculate it: it is about 3.8616Ć10ā33 m), we get what we should find:

This is a rather spectacular result, and one that would ā a priori ā support the interpretation of the wavefunction that is being suggested in this paper.

# VIII. The boson-fermion dichotomy

Let us do some more thinking on the boson-fermion dichotomy. Again, we should remind ourselves that an actual particle is localized in space and that it can, therefore, not be represented by the elementary wavefunction Ļ = aĀ·eāi[EĀ·t ā pāx]/Ä§ or, for a particle at rest, the Ļ = aĀ·eāiāEĀ·t/Ä§ function. We must build a wave packet for that: a sum of wavefunctions, each with their own amplitude ai, and their own Ļi = āEi/Ä§. Each of these wavefunctions will contribute some energy to the total energy of the wave packet. Now, we can have another wild but logical theory about this.

Think of the apparent right-handedness of the elementary wavefunction: surely, Nature can’t be bothered about our convention of measuring phase angles clockwise or counterclockwise. Also, the angular momentum can be positive or negative: J = +Ä§/2 or āÄ§/2. Hence, we would probably like to think that an actual particle – think of an electron, or whatever other particle you’d think of – may consist of right-handed as well as left-handed elementary waves. To be precise, we may think they either consist of (elementary) right-handed waves or, else, of (elementary) left-handed waves. An elementary right-handed wave would be written as:

Ļ(Īøi= aiĀ·(cosĪøi + iĀ·sinĪøi)

In contrast, an elementary left-handed wave would be written as:

Ļ(Īøi= aiĀ·(cosĪøi ā iĀ·sinĪøi)

How does that work out with the E0Ā·t argument of our wavefunction? Position is position, and direction is direction, but time? Time has only one direction, but Nature surely does not care how we count time: counting like 1, 2, 3, etcetera or like ā1, ā2, ā3, etcetera is just the same. If we count like 1, 2, 3, etcetera, then we write our wavefunction like:

Ļ = aĀ·cos(E0āt/Ä§) ā iĀ·aĀ·sin(E0āt/Ä§)

If we count time like ā1, ā2, ā3, etcetera then we write it as:

Ļ = aĀ·cos(āE0āt/Ä§) ā iĀ·aĀ·sin(āE0āt/Ä§)= aĀ·cos(E0āt/Ä§) + iĀ·aĀ·sin(E0āt/Ä§)

Hence, it is just like the left- or right-handed circular polarization of an electromagnetic wave: we can have both for the matter-wave too! This, then, should explain why we can have either positive or negative quantum-mechanical spin (+Ä§/2 or āÄ§/2). It is the usual thing: we have two mathematical possibilities here, and so we must have two physical situations that correspond to it.

It is only natural. If we have left- and right-handed photons – or, generalizing, left- and right-handed bosons – then we should also have left- and right-handed fermions (electrons, protons, etcetera). Back to the dichotomy. The textbook analysis of the dichotomy between bosons and fermions may be epitomized by Richard Feynmanās Lecture on it (Feynman, III-4), which is confusing and ā I would dare to say ā even inconsistent: how are photons or electrons supposed to know that they need to interfere with a positive or a negative sign? They are not supposed to know anything: knowledge is part of our interpretation of whatever it is that is going on there.

Hence, it is probably best to keep it simple, and think of the dichotomy in terms of the different physical dimensions of the oscillation: newton per kg versus newton per coulomb. And then, of course, we should also note that matter-particles have a rest mass and, therefore, actually carry charge. Photons do not. But both are two-dimensional oscillations, and the point is: the so-called vacuum – and the rest mass of our particle (which is zero for the photon and non-zero for everything else) – give us the natural frequency for both oscillations, which is beautifully summed up in that remarkable equation for the group and phase velocity of the wavefunction, which applies to photons as well as matter-particles:

(vphaseĀ·c)Ā·(vgroupĀ·c) = 1 ā vpĀ·vg = c2

The final question then is: why are photons spin-zero particles? Well… We should first remind ourselves of the fact that they do have spin when circularly polarized.[25] Here we may think of the rotation of the equivalent mass of their energy. However, if they are linearly polarized, then there is no spin. Even for circularly polarized waves, the spin angular momentum of photons is a weird concept. If photons have no (rest) mass, then they cannot carry any charge. They should, therefore, not have any magnetic moment. Indeed, what I wrote above shows an explanation of quantum-mechanical spin requires both mass as well as charge.[26]

# IX. Concluding remarks

There are, of course, other ways to look at the matter ā literally. For example, we can imagine two-dimensional oscillations as circular rather than linear oscillations. Think of a tiny ball, whose center of mass stays where it is, as depicted below. Any rotation ā around any axis ā will be some combination of a rotation around the two other axes. Hence, we may want to think of a two-dimensional oscillation as an oscillation of a polar and azimuthal angle.

Figure 8: Two-dimensional circular movement

The point of this paper is not to make any definite statements. That would be foolish. Its objective is just to challenge the simplistic mainstream viewpoint on the reality of the wavefunction. Stating that it is a mathematical construct only without physical significance amounts to saying it has no meaning at all. That is, clearly, a non-sustainable proposition.

The interpretation that is offered here looks at amplitude waves as traveling fields. Their physical dimension may be expressed in force per mass unit, as opposed to electromagnetic waves, whose amplitudes are expressed in force per (electric) charge unit. Also, the amplitudes of matter-waves incorporate a phase factor, but this may actually explain the rather enigmatic dichotomy between fermions and bosons and is, therefore, an added bonus.

The interpretation that is offered here has some advantages over other explanations, as it explains the how of diffraction and interference. However, while it offers a great explanation of the wave nature of matter, it does not explain its particle nature: while we think of the energy as being spread out, we will still observe electrons and photons as pointlike particles once they hit the detector. Why is it that a detector can sort of āhookā the whole blob of energy, so to speak?

The interpretation of the wavefunction that is offered here does not explain this. Hence, the complementarity principle of the Copenhagen interpretation of the wavefunction surely remains relevant.

# Appendix 1: The de Broglie relations and energy

The 1/2 factor in SchrĆ¶dingerās equation is related to the concept of the effective mass (meff). It is easy to make the wrong calculations. For example, when playing with the famous de Broglie relations ā aka as the matter-wave equations ā one may be tempted to derive the following energy concept:

1. E = hĀ·f and p = h/Ī». Therefore, f = E/h and Ī» = p/h.
2. v = fĀ·Ī» = (E/h)ā(p/h) = E/p
3. p = mĀ·v. Therefore, E = vĀ·p = mĀ·v2

E = mĀ·v2? This resembles the E = mc2 equation and, therefore, one may be enthused by the discovery, especially because the mĀ·v2 also pops up when working with the Least Action Principle in classical mechanics, which states that the path that is followed by a particle will minimize the following integral:Now, we can choose any reference point for the potential energy but, to reflect the energy conservation law, we can select a reference point that ensures the sum of the kinetic and the potential energy is zero throughout the time interval. If the force field is uniform, then the integrand will, effectively, be equal to KE ā PE = mĀ·v2.[27]

However, that is classical mechanics and, therefore, not so relevant in the context of the de Broglie equations, and the apparent paradox should be solved by distinguishing between the group and the phase velocity of the matter wave.

# Appendix 2: The concept of the effective mass

The effective mass ā as used in SchrĆ¶dingerās equation ā is a rather enigmatic concept. To make sure we are making the right analysis here, I should start by noting you will usually see SchrĆ¶dingerās equation written as:This formulation includes a term with the potential energy (U). In free space (no potential), this term disappears, and the equation can be re-written as:

āĻ(x, t)/āt = iĀ·(1/2)Ā·(Ä§/meff)Ā·ā2Ļ(x, t)

We just moved the iĀ·Ä§ coefficient to the other side, noting that 1/i = –i. Now, in one-dimensional space, and assuming Ļ is just the elementary wavefunction (so we substitute aĀ·eāiā[EĀ·t ā pāx]/Ä§ for Ļ), this implies the following:

āaĀ·iĀ·(E/Ä§)Ā·eāiā[EĀ·t ā pāx]/Ä§ = āiĀ·(Ä§/2meff)Ā·aĀ·(p2/Ä§2)Ā· eāiā[EĀ·t ā pāx]/Ä§

ā E = p2/(2meff) ā meff = mā(v/c)2/2 = māĪ²2/2

It is an ugly formula: it resembles the kinetic energy formula (K.E. = māv2/2) but it is, in fact, something completely different. The Ī²2/2 factor ensures the effective mass is always a fraction of the mass itself. To get rid of the ugly 1/2 factor, we may re-define meff as two times the old meff (hence, meffNEW = 2āmeffOLD), as a result of which the formula will look somewhat better:

meff = mā(v/c)2 = māĪ²2

We know Ī² varies between 0 and 1 and, therefore, meff will vary between 0 and m. Feynman drops the subscript, and just writes meff as m in his textbook (see Feynman, III-19). On the other hand, the electron mass as used is also the electron mass that is used to calculate the size of an atom (see Feynman, III-2-4). As such, the two mass concepts are, effectively, mutually compatible. It is confusing because the same mass is often defined as the mass of a stationary electron (see, for example, the article on it in the online Wikipedia encyclopedia[28]).

In the context of the derivation of the electron orbitals, we do have the potential energy term ā which is the equivalent of a source term in a diffusion equation ā and that may explain why the above-mentioned meff = mā(v/c)2 = māĪ²2 formula does not apply.

# References

This paper discusses general principles in physics only. Hence, references can be limited to references to physics textbooks only. For ease of reading, any reference to additional material has been limited to a more popular undergrad textbook that can be consulted online: Feynmanās Lectures on Physics (http://www.feynmanlectures.caltech.edu). References are per volume, per chapter and per section. For example, Feynman III-19-3 refers to Volume III, Chapter 19, Section 3.

# Notes

[1] Of course, an actual particle is localized in space and can, therefore, not be represented by the elementary wavefunction Ļ = aĀ·eāiāĪøaĀ·eāi[EĀ·t ā pāx]/Ä§ = aĀ·(cosĪø iĀ·aĀ·sinĪø). We must build a wave packet for that: a sum of wavefunctions, each with its own amplitude ak and its own argument Īøk = (Ekāt – pkāx)/Ä§. This is dealt with in this paper as part of the discussion on the mathematical and physical interpretation of the normalization condition.

[2] The N/kg dimension immediately, and naturally, reduces to the dimension of acceleration (m/s2), thereby facilitating a direct interpretation in terms of Newtonās force law.

[3] In physics, a two-spring metaphor is more common. Hence, the pistons in the authorās perpetuum mobile may be replaced by springs.

[4] The author re-derives the equation for the Compton scattering radius in section VII of the paper.

[5] The magnetic force can be analyzed as a relativistic effect (see Feynman II-13-6). The dichotomy between the electric force as a polar vector and the magnetic force as an axial vector disappears in the relativistic four-vector representation of electromagnetism.

[6] For example, when using SchrĆ¶dingerās equation in a central field (think of the electron around a proton), the use of polar coordinates is recommended, as it ensures the symmetry of the Hamiltonian under all rotations (see Feynman III-19-3)

[7] This sentiment is usually summed up in the apocryphal quote: āGod does not play dice.āThe actual quote comes out of one of Einsteinās private letters to Cornelius Lanczos, another scientist who had also emigrated to the US. The full quote is as follows: “You are the only person I know who has the same attitude towards physics as I have: belief in the comprehension of reality through something basically simple and unified… It seems hard to sneak a look at God’s cards. But that He plays dice and uses ‘telepathic’ methods… is something that I cannot believe for a single moment.” (Helen Dukas and Banesh Hoffman, Albert Einstein, the Human Side: New Glimpses from His Archives, 1979)

[8] Of course, both are different velocities: Ļ is an angular velocity, while v is a linear velocity: Ļ is measured in radians per second, while v is measured in meter per second. However, the definition of a radian implies radians are measured in distance units. Hence, the physical dimensions are, effectively, the same. As for the formula for the total energy of an oscillator, we should actually write: E = mĀ·a2āĻ2/2. The additional factor (a) is the (maximum) amplitude of the oscillator.

[9] We also have a 1/2 factor in the E = mv2/2 formula. Two remarks may be made here. First, it may be noted this is a non-relativistic formula and, more importantly, incorporates kinetic energy only. Using the Lorentz factor (Ī³), we can write the relativistically correct formula for the kinetic energy as K.E. = E ā E0 = mvc2 ā m0c2 = m0Ī³c2 ā m0c2 = m0c2(Ī³ ā 1). As for the exclusion of the potential energy, we may note that we may choose our reference point for the potential energy such that the kinetic and potential energy mirror each other. The energy concept that then emerges is the one that is used in the context of the Principle of Least Action: it equals E = mv2. Appendix 1 provides some notes on that.

[10] Instead of two cylinders with pistons, one may also think of connecting two springs with a crankshaft.

[11] It is interesting to note that we may look at the energy in the rotating flywheel as potential energy because it is energy that is associated with motion, albeit circular motion. In physics, one may associate a rotating object with kinetic energy using the rotational equivalent of mass and linear velocity, i.e. rotational inertia (I) and angular velocity Ļ. The kinetic energy of a rotating object is then given by K.E. = (1/2)Ā·IĀ·Ļ2.

[12] Because of the sideways motion of the connecting rods, the sinusoidal function will describe the linear motion only approximately, but you can easily imagine the idealized limit situation.

[13] The Ļ2= 1/LC formula gives us the natural or resonant frequency for a electric circuit consisting of a resistor (R), an inductor (L), and a capacitor (C). Writing the formula as Ļ2= C1/L introduces the concept of elastance, which is the equivalent of the mechanical stiffness (k) of a spring.

[14] The resistance in an electric circuit introduces a damping factor. When analyzing a mechanical spring, one may also want to introduce a drag coefficient. Both are usually defined as a fraction of the inertia, which is the mass for a spring and the inductance for an electric circuit. Hence, we would write the resistance for a spring as Ī³m and as R = Ī³L respectively.

[15] Photons are emitted by atomic oscillators: atoms going from one state (energy level) to another. Feynman (Lectures, I-33-3) shows us how to calculate the Q of these atomic oscillators: it is of the order of 108, which means the wave train will last about 10ā8 seconds (to be precise, that is the time it takes for the radiation to die out by a factor 1/e). For example, for sodium light, the radiation will last about 3.2Ć10ā8 seconds (this is the so-called decay time Ļ). Now, because the frequency of sodium light is some 500 THz (500Ć1012 oscillations per second), this makes for some 16 million oscillations. There is an interesting paradox here: the speed of light tells us that such wave train will have a length of about 9.6 m! How is that to be reconciled with the pointlike nature of a photon? The paradox can only be explained by relativistic length contraction: in an analysis like this, one need to distinguish the reference frame of the photon ā riding along the wave as it is being emitted, so to speak ā and our stationary reference frame, which is that of the emitting atom.

[16] This is a general result and is reflected in the K.E. = T = (1/2)Ā·mĀ·Ļ2Ā·a2Ā·sin2(ĻĀ·t + Ī) and the P.E. = U = kĀ·x2/2 = (1/2)Ā· mĀ·Ļ2Ā·a2Ā·cos2(ĻĀ·t + Ī) formulas for the linear oscillator.

[17] Feynman further formalizes this in his Lecture on Superconductivity (Feynman, III-21-2), in which he refers to SchrĆ¶dingerās equation as the āequation for continuity of probabilitiesā. The analysis is centered on the local conservation of energy, which confirms the interpretation of SchrĆ¶dingerās equation as an energy diffusion equation.

[18] The meff is the effective mass of the particle, which depends on the medium. For example, an electron traveling in a solid (a transistor, for example) will have a different effective mass than in an atom. In free space, we can drop the subscript and just write meff = m. Appendix 2 provides some additional notes on the concept. As for the equations, they are easily derived from noting that two complex numbers a + iāb and c + iād are equal if, and only if, their real and imaginary parts are the same. Now, the āĻ/āt = iā(Ä§/meff)āā2Ļ equation amounts to writing something like this: a + iāb = iā(c + iād). Now, remembering that i2 = ā1, you can easily figure out that iā(c + iād) = iāc + i2ād = ā d + iāc.

[19] The dimension of B is usually written as N/(māA), using the SI unit for current, i.e. the ampere (A). However, 1 C = 1 Aās and, hence, 1 N/(māA) = 1 (N/C)/(m/s).

[20] Of course, multiplication with i amounts to a counterclockwise rotation. Hence, multiplication by –i also amounts to a rotation by 90 degrees, but clockwise. Now, to uniquely identify the clockwise and counterclockwise directions, we need to establish the equivalent of the right-hand rule for a proper geometric interpretation of SchrĆ¶dingerās equation in three-dimensional space: if we look at a clock from the back, then its hand will be moving counterclockwise. When writing B = (1/c)āiāE, we assume we are looking in the negative x-direction. If we are looking in the positive x-direction, we should write: B = -(1/c)āiāE. Of course, Nature does not care about our conventions. Hence, both should give the same results in calculations. We will show in a moment they do.

[21] In fact, when multiplying C2/(NĀ·m2) with N2/C2, we get N/m2, but we can multiply this with 1 = m/m to get the desired result. It is significant that an energy density (joule per unit volume) can also be measured in newton (force per unit area.

[22] The illustration shows a linearly polarized wave, but the obtained result is general.

[23] The sine and cosine are essentially the same functions, except for the difference in the phase: sinĪø = cos(ĪøāĻ /2).

[24] I must thank a physics blogger for re-writing the 1/(Īµ0Ā·Ī¼0) = c2 equation like this. See: http://reciprocal.systems/phpBB3/viewtopic.php?t=236 (retrieved on 29 September 2017).

[25] A circularly polarized electromagnetic wave may be analyzed as consisting of two perpendicular electromagnetic plane waves of equal amplitude and 90Ā° difference in phase.

[26] Of course, the reader will now wonder: what about neutrons? How to explain neutron spin? Neutrons are neutral. That is correct, but neutrons are not elementary: they consist of (charged) quarks. Hence, neutron spin can (or should) be explained by the spin of the underlying quarks.

[27] We detailed the mathematical framework and detailed calculations in the following online article: https://readingfeynman.org/2017/09/15/the-principle-of-least-action-re-visited.

[28] https://en.wikipedia.org/wiki/Electron_rest_mass (retrieved on 29 September 2017).

# Playing with amplitudes

Let’s play a bit with the stuff we found in our previous post. This is going to be unconventional, or experimental, if you want. The idea is to give you… Well… Some ideas. So you can play yourself. š Let’s go.

Let’s first look at Feynman’s (simplified) formula for the amplitude of a photon to go from point aĀ to point b. If we identify point aĀ by the position vector r1Ā and point bĀ by the position vectorĀ r2, and using Dirac’s fancyĀ bra-ketĀ notation, then it’s written as:

So we have a vector dot product here: pār12Ā = |p|ā|r12|Ā·Ā cosĪø = pār12Ā·cosĪ±. The angle here (Ī±) is the angle between theĀ pĀ andĀ r12Ā vector. All good. Well… No. We’ve got a problem. When it comes to calculating probabilities, the Ī± angle doesn’t matter: |eiĀ·Īø/r|2Ā = 1/r2. Hence, for the probability, we get: P = |Ā ā©r2|r1āŖ |2Ā =Ā 1/r122. Always ! Now that’s strange. The Īø =Ā pār12/Ä§Ā argument gives us a different phase depending on the angle (Ī±) between p and r12. But… Well… Think of it:Ā cosĪ± goes from 1 to 0 when Ī± goes from 0 to Ā±90Ā° and, of course, is negative when p and r12Ā have opposite directions but… Well… According to this formula, the probabilitiesĀ doĀ not depend on the direction of the momentum. That’s just weird, I think. Did Feynman, in his iconicĀ Lectures, give us a meaningless formula?

Maybe. We may also note this function looks like the elementary wavefunction for any particle, which we wrote as:

Ļ(x, t) = aĀ·eāiāĪøĀ =Ā aĀ·eāiā(Eāt āĀ pāx)/Ä§= aĀ·eāiā(Eāt)/Ä§Ā·eiā(pāx)/Ä§

The only difference is that the ā©r2|r1āŖ sort of abstracts away from time, so… Well… Let’s get a feel for the quantities. Let’s think of a photon carryingĀ some typical amount of energy. Hence, let’s talk visible light and, therefore, photons of a few eV only – say 5.625 eV = 5.625Ć1.6Ć10ā19Ā J = 9Ć10ā19Ā J. Hence, their momentum is equal to p = E/c = (9Ć10ā19Ā NĀ·m)/(3Ć105Ā m/s) = 3Ć10ā24Ā NĀ·s. That’s tiny but that’s only becauseĀ newtonsĀ andĀ secondsĀ are enormous units at the (sub-)atomic scale. As for the distance, we may want to use the thickness of a playing card as a starter, as that’s what Young used when establishing the experimentalĀ fact of light interfering with itself. Now, playing cards in Young’s time were obviously rougher than those today, but let’s take the smaller distance: modern cards are as thin as 0.3 mm. Still, that distance is associated with a value ofĀ Īø that is equal to 13.6 million. Hence, theĀ densityĀ of our wavefunction is enormous at this scale, and it’s a bit of a miracle that Young could see any interference at all ! As shown in the table below, we only get meaningful values (remember:Ā Īø is a phase angle) when we go down to the nanometerĀ scale (10ā9Ā m) or, even better, theĀ angstroms scale ((10ā9Ā m).Ā

So… Well… Again: what can we do with Feynman’s formula? Perhaps he didn’t give us a propagatorĀ function but something that is more general (read: more meaningful) at our (limited) level of knowledge. As I’ve been reading Feynman for quite a while now – like three or four years š – I think… Well… Yes. That’s it. Feynman wants us to think about it. š Are you joking again, Mr. Feynman?Ā šĀ So let’s assume the reasonable thing: let’s assume it gives us the amplitude to go from point a toĀ point bĀ by the position vectorĀ along some path r.Ā So, then, in line with what we wrote in our previous post, let’s say pĀ·rĀ (momentum over a distance) is the action (S) we’d associate with this particular path (r) and then see where we get. So let’s writeĀ the formula like this:

ĻĀ =Ā aĀ·eiĀ·ĪøĀ = (1/r)Ā·eiĀ·S/Ä§Ā =Ā eiĀ·pār/Ä§/r

We’ll use an index to denote the various paths: r0Ā is the straight-line path and riĀ is any (other) path.Ā Now, quantum mechanics tells us we should calculate this amplitudeĀ for every possible path. The illustration below shows the straight-line path and two nearby paths. So each of these paths is associated with some amount of action, which we measure in PlanckĀ units:Ā Īø =Ā S/Ä§.Ā

The time interval is given by tĀ = t0Ā =Ā r0/c, for all paths. Why is the time interval the same for all paths? Because we think of a photon going from some specificĀ point in space and in timeĀ to some otherĀ specificĀ point in space and in time. Indeed, when everything is said and done, we do think of light as traveling from pointĀ a to pointĀ bĀ at the speed of light (c). In fact, all of the weird stuff here is all about trying to explain howĀ it does that. š

Now, if we would think of the photon actually traveling along this or that path, then this implies its velocityĀ along any of the nonlinear paths will be largerĀ thanĀ c, which is OK. That’s just the weirdness of quantum mechanics, and you should actuallyĀ notĀ think of the photon actually traveling along one of these paths anyway although we’ll often put it that way. Think of something fuzzier, whatever that may be. š

So the action is energy times time, or momentum times distance. Hence, the difference in action between two paths iĀ andĀ jĀ is given by:

Ī“SĀ = pĀ·rjĀ āĀ pĀ·riĀ = pĀ·(rjĀ ā ri) = pĀ·Īr

I’ll explain theĀ Ī“S <Ā 2ĻÄ§/3 thing in a moment. Let’s first pause and think about theĀ uncertainty and how we’re modeling it. We can effectively think of the variation in SĀ as some uncertaintyĀ in the action: Ī“SĀ = ĪS = pĀ·Īr. However, if SĀ is also equal to energy times time (SĀ = EĀ·t), and we insist tĀ is the same for all paths, then we must have some uncertainty in the energy, right? Hence, we can write Ī“SĀ as ĪS = ĪEĀ·t. But, of course, E =Ā E =Ā mĀ·c2Ā = pĀ·c, so we will have an uncertainty in the momentum as well. Hence, the variation inĀ SĀ should be written as:

Ī“SĀ = ĪSĀ = ĪpĀ·Īr

That’s just logical thinking: if we, somehow, entertain the idea of a photon going from someĀ specificĀ point in spacetime to some otherĀ specificĀ point in spacetime along various paths, then the variation, or uncertainty,Ā in the action will effectively combine some uncertainty in the momentum and the distance. We can calculate Īp as ĪE/c, so we get the following:

Ī“SĀ = ĪSĀ = ĪpĀ·Īr =Ā ĪEĀ·Īr/c = ĪEĀ·Īt with Īt =Ā Īr/c

So we have the two expressions for the Uncertainty Principle here: ĪSĀ = ĪpĀ·Īr =Ā ĪEĀ·Īt. Just be careful with the interpretation of Īt: it’s just the equivalent of Īr. We just express the uncertainty in distance in secondsĀ using the (absolute) speed of light. We are notĀ changing our spacetime interval: we’re still looking at a photon going fromĀ aĀ toĀ bĀ inĀ tĀ seconds,Ā exactly. Let’s now look at theĀ Ī“S <Ā 2ĻÄ§/3 thing. If we’re addingĀ twoĀ amplitudes (twoĀ arrowsĀ or vectors, so to speak) and we want the magnitude of the result to be larger than the magnitude of the two contributions, then the angle between them should be smaller than 120 degrees, so that’s 2Ļ/3 rad. The illustration below shows how you can figure that out geometrically.Hence, if S0Ā is the action for r0, then S1Ā = S0Ā + Ä§Ā and S2Ā = S0Ā + 2Ā·Ä§ are still good, but S3Ā = S0Ā + 3Ā·Ä§Ā isĀ notĀ good. Why? Because the difference in the phase angles is ĪĪøĀ =Ā S1/Ä§Ā āĀ S0/Ä§Ā = (S0Ā + Ä§)/Ä§Ā āĀ S0/Ä§ = 1 andĀ ĪĪø =Ā S2/Ä§Ā āĀ S0/Ä§Ā = (S0Ā + 2Ā·Ä§)/Ä§Ā āĀ S0/Ä§ = 2 respectively, so that’s 57.3Ā°Ā and 114.6Ā°Ā respectively and that’s, effectively,Ā lessĀ than 120Ā°. In contrast,Ā for the next path, we find that ĪĪøĀ =Ā S3/Ä§Ā āĀ S0/Ä§Ā = (S0Ā + 3Ā·Ä§)/Ä§Ā āĀ S0/Ä§ = 3, so that’s 171.9Ā°. So that amplitude gives us a negative contribution.

Let’s do some calculations using a spreadsheet. To simplify things, we will assume we measure everything (time, distance, force, mass, energy, action,…) in Planck units. Hence, we can simply write:Ā SnĀ = S0Ā + n. Of course, nĀ = 1, 2,… etcetera, right? Well… Maybe not. We areĀ measuringĀ action in units ofĀ Ä§, butĀ do we actually think actionĀ comesĀ in units ofĀ Ä§?Ā I am not sure. It would make sense, intuitively, butā¦ Wellā¦ Thereās uncertainty on the energy (E) and the momentum (p) of our photon, right? And how accurately can we measure the distance? So thereās some randomness everywhere. š¦ So let’s leave that question open as for now.

We will also assume that the phase angle forĀ S0Ā is equal to 0 (or some multiple of 2Ļ, if you want). That’s just a matter of choosing the origin of time. This makes it really easy: ĪSnĀ =Ā SnĀ ā S0Ā = n, and the associated phase angle ĪønĀ = ĪĪønĀ is the same. In short, the amplitude for each path reduces to ĻnĀ = eiĀ·n/r0. So we need to add these firstĀ andĀ thenĀ calculate the magnitude, which we can then square to get a probability. Of course, there is also the issue of normalization (probabilities have to add up to one) but let’s tackle that later. For the calculations, we use Euler’s rĀ·eiĀ·ĪøĀ = rĀ·(cosĪø + iĀ·sinĪø) = rĀ·cosĪø + iĀ·rĀ·sinĪø formula. Needless to say, |rĀ·eiĀ·Īø|2Ā = |r|2Ā·|eiĀ·Īø|2Ā = |r|2Ā·(cos2Īø + sin2Īø) = r. Finally, when adding complex numbers, we add the real and imaginary parts respectively, and we’ll denote the Ļ0Ā + Ļ1Ā +Ļ2Ā + … sum as ĪØ.

Now, we also need to see how our ĪSĀ = ĪpĀ·ĪrĀ works out. We may want to assume that the uncertainty in p and in r will both be proportional to the overall uncertainty in the action. For example, we could try writing the following:Ā ĪSnĀ = ĪpnĀ·ĪrnĀ =Ā nĀ·Īp1Ā·Īr1. It also makes sense that you may want ĪpnĀ and ĪrnĀ to be proportional to Īp1Ā and Īr1Ā respectively. Combining both, the assumption would be this:

ĪpnĀ =Ā ānĀ·Īp1Ā andĀ ĪrnĀ =Ā ānĀ·Īr1

So now we just need to decide how we will distribute ĪS1Ā =Ā Ä§Ā = 1 over Īp1Ā and Īr1Ā respectively. For example, if we’d assume Īp1Ā = 1, then Īr1Ā = Ä§/Īp1Ā = 1/1 = 1. These are the calculations. I will let you analyze them. šWell… We get a weird result. It reminds me ofĀ Feynman’s explanation of the partial reflection of light, shown below, but… Well… That doesn’t make much sense, does it?

Hmm… Maybe it does. š Look at the graph more carefully. The peaks sort of oscillate out so… Well… That might make sense… š

Does it? Are we doingĀ something wrongĀ here? These amplitudes should reflect the ones that are reflected in those nice animations (like this one, for example, which is part of thatās part of the Wikipedia article on Feynmanās path integral formulation of quantum mechanics). So what’s wrong, if anything? Well… Our paths differ by some fixed amount of action, which doesn’t quite reflect the geometric approach that’s used in those animations. The graph below shows how the distanceĀ rĀ varies as a function ofĀ n.Ā

If we’d use a model in which the distance wouldĀ increaseĀ linearly or, preferably, exponentially, then we’d get the result we want to get, right?

Well… Maybe. Let’s try it.Ā Hmm… We need to think about the geometry here. Look at the triangle below.Ā IfĀ bĀ is the straight-line path (r0), thenĀ acĀ could be one of the crooked paths (rn). To simplify, we’ll assume isosceles triangles, soĀ aĀ equalsĀ cĀ and, hence, rnĀ = 2Ā·a = 2Ā·c. We will also assume theĀ successive paths are separated by the same vertical distance (h =Ā h1) right in the middle, so hbĀ =Ā hnĀ = nĀ·h1.Ā It is then easy to show the following:This gives the following graph for rnĀ = 10 and h1Ā = 0.01.

Is this the right step increase? Not sure. We can vary the values in our spreadsheet. Let’s first build it. TheĀ photon will have to travel faster in order to cover the extra distance in the same time, so its momentum will be higher. Let’s think about the velocity. Let’s start with the first path (nĀ = 1). In order to cover the extraĀ distance Īr1, the velocity c1Ā must be equal to (r0Ā + Īr1)/tĀ = r0/tĀ + Īr1/t =Ā cĀ + Īr1/tĀ = c0Ā + Īr1/t. We can write c1Ā as c1Ā =Ā c0Ā + Īc1, so Īc1Ā = Īr1/t.Ā Now, theĀ ratioĀ of p1Ā  and p0Ā will be equal to theĀ ratioĀ of c1Ā andĀ c0Ā because p1/p0Ā = (mc1)/mc0) = c1/c0. Hence, we have the following formula for p1:

p1Ā = p0Ā·c1/c0Ā = p0Ā·(c0Ā + Īc1)/c0Ā = p0Ā·[1 + Īr1/(c0Ā·t) = p0Ā·(1 + Īr1/r0)

ForĀ pn, the logic is the same, so we write:

pnĀ = p0Ā·cn/c0Ā = p0Ā·(c0Ā + Īcn)/c0Ā = p0Ā·[1 + Īrn/(c0Ā·t) = p0Ā·(1 + Īrn/r0)

Let’s do the calculations, and let’s use meaningful values, so the nanometer scale and actual values for Planck’s constant and the photon momentum. The results are shown below.Ā

Pretty interesting. In fact, this looksĀ reallyĀ good. TheĀ probabilityĀ first swings around wildly, because of these zones of constructive and destructive interference, but then stabilizes. [Of course, I would need to normalize the probabilities, but you get the idea, right?] So… Well… I think we get a veryĀ meaningful result with this model. Sweet ! š I’m lovin’ it ! š And, here you go, this is (part of) the calculation table, so you can see what I am doing. š

The graphs below look even better: I just changed the h1/r0Ā ratio from 1/100 to 1/10. The probability stabilizes almost immediately. š So… Well… It’s not as fancy as the referenced animation, but I think the educational value of this thing here is at least as good ! š

š This is good stuff… š

Post scriptum (19 September 2017): There is an obvious inconsistency in the model above, and in the calculations. We assume there is a path r1Ā = ,Ā r2, r2,etcetera, and then we calculate the action for it, and the amplitude, and then we add the amplitude to the sum. But, surely, we should count these paths twice, in two-dimensional space, that is. Think of the graph: we have positive and negative interference zones that are sort of layered around the straight-line path, as shown below.

In three-dimensional space, these lines become surfaces. Hence, rather than adding oneĀ arrow for everyĀ Ī“Ā Ā having oneĀ contribution only, we may want to add… Well… In three-dimensional space, the formula for the surface around the straight-line path would probably look like ĻĀ·hnĀ·r1, right? Hmm… Interesting idea. I changed my spreadsheet to incorporate that idea, and I got the graph below. It’s a nonsensical result, because the probability does swing around, but it gradually spins out of control: it never stabilizes.That’s because we increase theĀ weightĀ of the paths that are further removed from the center. So… Well… We shouldn’t be doing that, I guess. š I’ll you look for the right formula, OK? Let me know when you found it. š

# Some thoughts on the nature of reality

Some other comment on an article on my other blog, inspired me to structure some thoughts that are spread over various blog posts. What follows below, is probably the first draft of an article or a paper I plan to write. Or, who knows, I might re-write my two introductory books on quantum physics and publish a new edition soon. š

## Physical dimensions and Uncertainty

The physical dimension of the quantum of action (h orĀ Ä§ = h/2Ļ) is force (expressed in newton)Ā times distance (expressed in meter)Ā times time (expressed in seconds): NĀ·mĀ·s. Now, you may think this NĀ·mĀ·s dimension is kinda hard to imagine. We can imagine its individual components, right? Force, distance and time. We know what they are. But the product of all three? What is it, really?

It shouldn’t be all that hard to imagine what it might be, right? The NĀ·mĀ·s unit is also the unit in which angular momentum is expressed – and you can sort of imagine what that is, right? Think of a spinning top, or a gyroscope. We may also think of the following:

1. [h] = NĀ·mĀ·s = (NĀ·m)Ā·s = [E]Ā·[t]
2. [h] = NĀ·mĀ·s = (NĀ·s)Ā·m = [p]Ā·[x]

Hence, the physical dimension of action is that of energy (E) multiplied by time (t) or, alternatively, that of momentum (p) times distance (x). To be precise, the second dimensional equation should be written as [h] = [p]Ā·[x], because both the momentum and the distance traveled will be associated with some direction. It’s a moot point for the discussion at the moment, though. Let’s think about the first equation first:Ā [h] = [E]Ā·[t]. What does it mean?

Energy… Hmm… InĀ real life, we are usually not interested in the energy of a system as such, but by the energy it can deliver, or absorb, per second. This is referred to as theĀ powerĀ of a system, and it’s expressed in J/s, or watt. Power is also defined as the (time) rate at which work is done. Hmm… But so here we’re multiplying energy and time. So what’s that? After Hiroshima and Nagasaki, we can sort of imagine the energy of an atomic bomb. We can also sort of imagine the power that’s being released by the Sun in light and other forms of radiation, which is about 385Ć1024 joule per second. But energy times time? What’s that?

I am not sure. If we think of the Sun as a huge reservoir of energy, then the physical dimension of action is just like having that reservoir of energy guaranteed for some time, regardless of how fast or how slow we use it. So, in short, it’s just like the Sun – or the Earth, or the Moon, or whatever object – just being there, for someĀ definiteĀ amount of time. So, yes: someĀ definite amount of mass or energy (E) for someĀ definiteĀ amount of time (t).

Let’s bring the mass-energy equivalence formula in here: E = mc2. Hence, the physical dimension of action can also be written as [h] = [E]Ā·[t] = [mc]2Ā·[t] = (kgĀ·m2/s2)Ā·s =Ā kgĀ·m2/s.Ā What does that say? Not all that much – for the time being, at least. We can get thisĀ [h] = kgĀ·m2/s through some other substitution as well. A force of one newton will give a mass of 1 kg an acceleration of 1 m/s per second. Therefore, 1 N = 1 kgĀ·m/s2Ā and, hence, the physical dimension of h, or the unit of angular momentum, may also be written as 1 NĀ·mĀ·s = 1 (kgĀ·m/s2)Ā·mĀ·s = 1 kgĀ·m2/s, i.e. the product of mass, velocity and distance.

Hmm… What can we do with that? Nothing much for the moment: our first reading of it is just that it reminds us of the definition of angular momentum – some mass with some velocity rotating around an axis. What about the distance? Oh… The distance here is just the distance from the axis, right? Right. But… Well… It’s like having some amount of linear momentum available over some distance – or in some space, right? That’s sufficiently significant as an interpretation for the moment, I’d think…

## Fundamental units

This makes one think about what units would be fundamental – and what units we’d consider as being derived. Formally, theĀ newton is aĀ derivedĀ unit in the metric system, as opposed to the units of mass, length and time (kg, m, s). Nevertheless, I personally like to think of force as being fundamental:Ā  a force is what causes an object to deviate from its straight trajectory in spacetime. Hence, we may want to think of theĀ quantum of action as representing three fundamental physical dimensions: (1)Ā force, (2)Ā time and (3) distance – or space. We may then look at energy and (linear) momentum as physical quantities combining (1) force and distance and (2) force and time respectively.

Let me write this out:

1. Force times length (think of a force that isĀ acting on some object over some distance) is energy: 1 jouleĀ (J) =Ā 1 newtonĀ·meter (N). Hence, we may think of the concept of energy as a projectionĀ of action in space only: we make abstraction of time. The physical dimension of the quantum of action should then be written as [h] = [E]Ā·[t]. [Note the square brackets tell us we are looking at aĀ dimensionalĀ equation only, so [t] is just the physical dimension of the time variable. It’s a bit confusing because I also use square brackets as parentheses.]
2. Conversely, the magnitude of linear momentum (p = mĀ·v) is expressed in newtonĀ·seconds: 1 kgĀ·m/s = 1 (kgĀ·m/s2)Ā·s = 1 NĀ·s. Hence, we may think of (linear) momentum as a projection of action in time only: we make abstraction of its spatial dimension. Think of a force that is acting on some objectĀ during some time.Ā The physical dimension of the quantum of action should then be written as [h] = [p]Ā·[x]

Of course, a force that is acting on some object during some time, will usually also act on the same object over some distance but… Well… Just try, for once, to make abstraction of one of the two dimensions here: timeĀ orĀ distance.

It is a difficult thing to do because, when everything is said and done, we don’t live in space or in time alone, but in spacetime and, hence, such abstractions are not easy. [Of course, now you’ll say that it’s easy to think of something that moves in time only: an object that is standing still does just that – but then we know movement is relative, so there is no such thing as an object that is standing still in spaceĀ in an absolute sense: Hence, objects never stand still in spacetime.] In any case, we should try such abstractions, if only because of the principle of least actionĀ is so essential and deep in physics:

1. In classical physics, the path of some object in a force field will minimizeĀ the total action (which is usually written as S) along that path.
2. In quantum mechanics, the same action integral will give us various values S – each corresponding to a particular path – and each path (and, therefore, each value of S, really) will be associated with a probability amplitude that will be proportional to some constant times eāiĀ·ĪøĀ =Ā eiĀ·(S/Ä§). Because Ä§ is so tiny, even a small change in S will give a completely different phase angle Īø. Therefore, most amplitudes will cancel each other out as we take the sum of the amplitudes over all possible paths: only the paths that nearlyĀ give the same phase matter. In practice, these are the paths that are associated with a variation in S of an order of magnitude that is equal to Ä§.

The paragraph above summarizes, in essence, Feynman’s path integral formulation of quantum mechanics. We may, therefore, think of the quantum of actionĀ expressingĀ itself (1) in time only, (2) in space only, or – much more likely – (3) expressing itself in both dimensions at the same time. Hence, if the quantum of action gives us the order of magnitudeĀ of the uncertainty – think of writing something like S Ā± Ä§, we may re-write our dimensional [Ä§] = [E]Ā·[t] and [Ä§] = [p]Ā·[x] equations as the uncertainty equations:

• ĪEĀ·Īt = Ä§Ā
• ĪpĀ·Īx = Ä§

You should note here that it is best to think of the uncertainty relations as aĀ pairĀ of equations, if only because you should also think of the concept of energy and momentum as representing different aspectsĀ of the same reality, as evidenced by the (relativistic) energy-momentum relation (E2Ā = p2c2Ā ā m02c4). Also, as illustrated below, the actual path – or, to be more precise, what we might associate with the concept of the actual path – is likely to be some mix of Īx and Īt. If Īt is very small, then Īx will be very large. In order to move over such distance, our particle will require a larger energy, so ĪE will be large. Likewise, if Īt is very large, then Īx will be very small and, therefore, ĪE will be very small. You can also reason in terms of Īx, and talk about momentum rather than energy. You will arrive at the same conclusions: the ĪEĀ·Īt = h and ĪpĀ·Īx = hĀ relations represent two aspects of the same reality – or, at the very least, what we mightĀ thinkĀ of as reality.

Also think of the following: ifĀ ĪEĀ·Īt =Ā hĀ and ĪpĀ·Īx =Ā h, thenĀ ĪEĀ·Īt =Ā ĪpĀ·Īx and, therefore,Ā ĪE/Īp must be equal to Īx/Īt. Hence, theĀ ratioĀ of the uncertainty about x (the distance) and the uncertainty about t (the time) equals theĀ ratioĀ of the uncertainty about E (the energy) and the uncertainty about p (the momentum).

Of course, you will note that the actual uncertainty relations have a factor 1/2 in them. This may be explained by thinking of both negative as well as positive variations in space and in time.

We will obviously want to do some more thinking about those physical dimensions. The idea of a force implies the idea of some object – of some mass on which the force is acting. Hence, let’s think about the concept of mass now. But… Well… Mass and energy are supposed to be equivalent, right? So let’s look at the concept of energyĀ too.

## Action, energy and mass

What isĀ energy, really? InĀ real life, we are usually not interested in the energy of a system as such, but by the energy it can deliver, or absorb, per second. This is referred to as theĀ powerĀ of a system, and it’s expressed in J/s. However, in physics, we always talk energy – not power – so… Well… What is the energy of a system?

According to the de BroglieĀ and Einstein – and so many other eminent physicists, of course – we should not only think of the kinetic energy of its parts, but also of their potential energy, and their restĀ energy, and – for an atomic system – we may add some internal energy, which may be binding energy, or excitation energy (think of a hydrogen atom in an excited state, for example). A lot of stuff. š But, obviously, Einstein’s mass-equivalence formula comes to mind here, and summarizes it all:

E = mĀ·c2

The m in this formula refers to mass – not to meter, obviously. Stupid remark, of course… But… Well… What is energy, really? What is mass,Ā really? What’s thatĀ equivalenceĀ between mass and energy,Ā really?

I don’t have the definite answer to that question (otherwise I’d be famous), but… Well… I do think physicists and mathematicians should invest more in exploring some basic intuitions here. As I explained in several posts, it is very tempting to think of energy as some kind of two-dimensional oscillation of mass. A force over some distance will cause a mass to accelerate. This is reflected in theĀ dimensional analysis:

[E] = [m]Ā·[c2] = 1 kgĀ·m2/s2Ā = 1 kgĀ·m/s2Ā·m = 1 NĀ·m

The kg and m/s2Ā factors make this abundantly clear: m/s2Ā is the physical dimension of acceleration: (the change in) velocity per time unit.

Other formulas now come to mind, such as the Planck-Einstein relation: E = hĀ·f = ĻĀ·Ä§. We could also write: E = h/T. Needless to say, T = 1/fĀ is theĀ periodĀ of the oscillation. So we could say, for example, that the energy of some particle times the period of the oscillation gives us Planck’s constant again. What does that mean? Perhaps it’s easier to think of it the other way around: E/f = h = 6.626070040(81)Ć10ā34Ā JĀ·s. Now, fĀ is the number of oscillationsĀ per second. Let’s write it asĀ fĀ = n/s, so we get:

E/fĀ = E/(n/s) = EĀ·s/nĀ = 6.626070040(81)Ć10ā34Ā JĀ·s ā E/nĀ = 6.626070040(81)Ć10ā34Ā J

What an amazing result! Our wavicle – be it a photon or a matter-particle – will alwaysĀ packĀ 6.626070040(81)Ć10ā34Ā jouleĀ inĀ oneĀ oscillation, so that’s the numericalĀ value of Planck’s constant which, of course, depends on our fundamentalĀ units (i.e. kg, meter, second, etcetera in the SI system).

Of course, the obvious question is: what’s oneĀ oscillation? If it’s a wave packet, the oscillations may not have the same amplitude, and we may also not be able to define an exact period. In fact, we should expect the amplitude and duration of each oscillation to be slightly different, shouldn’t we? And then…

Well… What’s an oscillation? We’re used toĀ countingĀ them:Ā nĀ oscillations per second, so that’sĀ per time unit. How many do we have in total? We wrote about that in our posts on the shape and size of a photon. We know photons are emitted by atomic oscillators – or, to put it simply, just atoms going from one energy level to another. Feynman calculated the Q of these atomic oscillators: itās of the order of 108Ā (see hisĀ Lectures,Ā I-33-3: itās a wonderfully simple exercise, and one that really shows his greatness as a physics teacher), so… Well… This wave train will last about 10ā8Ā seconds (thatās the time it takes for the radiation to die out by a factor 1/e). To give a somewhat more precise example,Ā for sodium light, which has a frequency of 500 THz (500Ć1012Ā oscillations per second) and a wavelength of 600 nm (600Ć10ā9Ā meter), the radiation will lasts about 3.2Ć10ā8Ā seconds. [In fact, thatās the time it takes for the radiationās energy to die out by a factor 1/e, so(i.e. the so-called decay time Ļ), so the wavetrain will actually lastĀ longer, but so the amplitude becomes quite small after that time.]Ā So… Well… Thatās a very short time but… Still, taking into account the rather spectacular frequency (500 THz) of sodium light, that makes for some 16 million oscillations and, taking into the account the rather spectacular speed of light (3Ć108Ā m/s), that makes for a wave train with a length of, roughly,Ā 9.6 meter. Huh? 9.6 meter!? But a photon is supposed to be pointlike, isn’it it? It has no length, does it?

That’s where relativity helps us out: as I wrote in one of my posts, relativistic length contraction may explain the apparent paradox. Using the reference frame of the photonĀ – so if we’d be traveling at speed c,ā ridingā with the photon, so to say, as itās being emitted – then we’d āseeā the electromagnetic transient as itās being radiated into space.

However, while we can associate some massĀ with the energy of the photon, none of what I wrote above explains what the (rest) mass of a matter-particle could possibly be.Ā There is no real answer to that, I guess. You’ll think of the Higgs field now but… Then… Well. The Higgs field is a scalar field. Very simple: some number that’s associated with some position in spacetime. That doesn’t explain very much, does it? š¦ When everything is said and done, the scientists who, in 2013 only, got the Nobel Price for their theory on the Higgs mechanism, simply tell us mass is some number. That’s something we knew already, right? š

## The reality of the wavefunction

The wavefunction is, obviously, a mathematical construct: aĀ descriptionĀ of reality using a very specific language. What language? Mathematics, of course! Math may not be universal (aliens might not be able to decipher our mathematical models) but it’s pretty good as a globalĀ tool of communication, at least.

The realĀ question is: is the descriptionĀ accurate? Does it match reality and, if it does, howĀ goodĀ is the match? For example, the wavefunction for an electron in a hydrogen atom looks as follows:

Ļ(r, t) = eāiĀ·(E/Ä§)Ā·tĀ·f(r)

As I explained in previous posts (see, for example, my recent postĀ on reality and perception), theĀ f(r) function basically provides some envelope for the two-dimensional eāiĀ·ĪøĀ =Ā eāiĀ·(E/Ä§)Ā·tĀ = cosĪø + iĀ·sinĪøĀ oscillation, with rĀ = (x, y, z),Ā Īø = (E/Ä§)Ā·tĀ = ĻĀ·tĀ and Ļ = E/Ä§. So it presumes theĀ duration of each oscillation is some constant. Why? Well… Look at the formula: this thing has a constant frequency in time. It’s only the amplitude that is varying as a function of the rĀ = (x, y, z) coordinates. š So… Well… If each oscillation is to alwaysĀ packĀ 6.626070040(81)Ć10ā34Ā joule, but the amplitude of the oscillation varies from point to point, then… Well… We’ve got a problem. The wavefunction above is likely to be an approximation of reality only. š The associated energy is the same, but… Well… Reality is probablyĀ notĀ the nice geometrical shape we associate with those wavefunctions.

In addition, we should think of theĀ Uncertainty Principle: thereĀ mustĀ be some uncertainty in the energy of the photons when our hydrogen atom makes a transition from one energy level to another. But then… Well… If our photon packs something like 16 million oscillations, and the order of magnitude of the uncertainty is only of the order ofĀ hĀ (or Ä§ = h/2Ļ) which, as mentioned above, is the (average) energy of oneĀ oscillation only, then we don’t have much of a problem here, do we? š

Post scriptum: In previous posts, we offered some analogies – or metaphors – to a two-dimensional oscillation (remember the V-2 engine?). Perhaps it’s all relatively simple. If we have some tiny little ball of mass – and its center of mass has to stay where it is – then any rotation – around any axis – will be some combination of a rotation around ourĀ x- and z-axis – as shown below. Two axes only. So we may want to think of a two-dimensionalĀ oscillation as an oscillation of the polar and azimuthal angle. š

# Thinking again…

One of the comments on my other blog made me think I should, perhaps, write something on waves again. The animation below shows theĀ elementaryĀ wavefunctionĀ Ļ =Ā aĀ·eāiĪøĀ = Ļ =Ā aĀ·eāiĀ·ĪøĀ Ā = aĀ·eāi(ĻĀ·tākĀ·x)Ā = aĀ·eā(i/Ä§)Ā·(EĀ·tāpĀ·x)Ā .We know this elementary wavefunction cannotĀ represent a real-lifeĀ particle. Indeed, the aĀ·eāiĀ·ĪøĀ function implies the probability of finding the particle – an electron, a photon, or whatever – would be equal to P(x, t) = |Ļ(x, t)|2Ā = |aĀ·eā(i/Ä§)Ā·(EĀ·tāpĀ·x)|2Ā = |a|2Ā·|eā(i/Ä§)Ā·(EĀ·tāpĀ·x)|2Ā = |a|2Ā·12= a2Ā everywhere. Hence, the particle would be everywhere – and, therefore, nowhere really. We need to localize the wave – or build a wave packet. We can do so by introducing uncertainty: we then addĀ a potentially infinite number of these elementary wavefunctions with slightly different values for E and p, and various amplitudes a. Each of these amplitudes will then reflect theĀ contributionĀ to the composite wave, which – in three-dimensional space – we can write as:

Ļ(r, t) = eāiĀ·(E/Ä§)Ā·tĀ·f(r)

As I explained in previous posts (see, for example, my recent postĀ on reality and perception), theĀ f(r) function basically provides some envelope for the two-dimensional eāiĀ·ĪøĀ =Ā eāiĀ·(E/Ä§)Ā·tĀ = cosĪø + iĀ·sinĪøĀ oscillation, with rĀ = (x, y, z),Ā Īø = (E/Ä§)Ā·tĀ = ĻĀ·tĀ and Ļ = E/Ä§.

Note that it looks like the wave propagatesĀ from left to right – in theĀ positive direction of an axis which we may refer to as the x-axis. Also note this perception results from the fact that, naturally, we’d associate time with theĀ rotationĀ of that arrow at the center – i.e. with the motion in the illustration,Ā while the spatial dimensions are just what they are: linear spatial dimensions. [This point is, perhaps, somewhat less self-evident than you may think at first.]

Now, the axis which points upwards is usually referred to as the z-axis, and the third and final axis – which points towardsĀ us –Ā would then be the y-axis, obviously.Ā Unfortunately, this definition would violate the so-called right-hand rule for defining a proper reference frame: the figures below shows the two possibilities – a left-handed and a right-handed reference frame – and it’s the right-handed reference (i.e. the illustration on the right) which we have to use in order to correctly define all directions, including the direction ofĀ rotationĀ of the argument of the wavefunction.Hence, if we don’t change the direction of the y– and z-axes – so we keep defining the z-axis as the axis pointing upwards, and the y-axis as the axis pointing towardsĀ us – then the positive direction of the x-axis would actually be the direction from right to left, and we should say that the elementary wavefunction in the animation above seems to propagate in the negativeĀ x-direction. [Note that this left- or right-hand rule is quite astonishing: simply swapping the direction ofĀ oneĀ axis of a left-handed frame makes it right-handed, and vice versa.]

Note my language when I talk about the direction of propagation of our wave. I wrote: it looks like, or it seems toĀ go in this or that direction. And I mean that: there is no real travelingĀ here. At this point, you may want to review a post I wrote for my son, which explains the basic math behind waves, and in which I also explained the animation below.

Note how the peaks and troughs of this pulse seem to move leftwards, but the wave packet (or theĀ groupĀ or theĀ envelopeĀ of the waveāwhatever you want to call it) moves to the right. The point is: the pulse itself doesn’tĀ travel left or right. Think of the horizontal axis in the illustration above as an oscillating guitar string: each point on the string just moves up and down. Likewise, if our repeated pulse would represent a physical wave in water, for example, then the water just stays where it is: it just moves up and down. Likewise, if we shake up some rope, the rope is not going anywhere: we just started some motionĀ that is traveling down the rope.Ā In other words, the phase velocity is just a mathematical concept. The peaks and troughs that seem to be traveling are just mathematical points that are ātravelingā left or right. Thatās why thereās no limit on the phase velocity: it canĀ – and, according to quantum mechanics, actually willĀ –Ā exceed the speed of light. In contrast, the groupĀ velocity – which is the actual speed of the particle that is being represented by the wavefunction – may approachĀ – or, in the case of a massless photon, will actually equalĀ –Ā the speed of light, but will never exceedĀ it, and itsĀ directionĀ will, obviously, have aĀ physicalĀ significance as it is, effectively, the direction of travel of our particle – be it an electron, a photon (electromagnetic radiation), or whatever.

Hence, you should not think theĀ spinĀ of a particle – integer or half-integer – is somehow related to the direction of rotation of the argument of the elementary wavefunction. It isn’t: Nature doesn’t give a damn about our mathematical conventions, and that’s what the direction of rotation of the argument of that wavefunction is: just some mathematical convention. That’s why we write aĀ·eāi(ĻĀ·tākĀ·x)Ā rather thanĀ aĀ·ei(ĻĀ·t+kĀ·x)Ā orĀ aĀ·ei(ĻĀ·tākĀ·x): it’s just because of the right-hand rule for coordinate frames, and also because Euler defined the counter-clockwise direction as theĀ positive direction of an angle. There’s nothing more to it.

OK. That’s obvious. Let me now return to my interpretation of Einstein’s E = mĀ·c2Ā formula (see my previous posts on this). I noted that, in the reference frame of the particle itself (see my basics page), the elementary wavefunction aĀ·eā(i/Ä§)Ā·(EĀ·tāpĀ·x)Ā reduces to aĀ·eā(i/Ä§)Ā·(E’Ā·t’): the origin of the reference frame then coincides with (the center of) our particle itself, and the wavefunction only varies with the time in the inertial reference frame (i.e. the properĀ time t’), with the rest energy of the object (E’) as the time scale factor. How should we interpret this?

Well… Energy is force times distance, and force is defined as that what causes some massĀ toĀ accelerate. To be precise, theĀ newtonĀ – as the unit of force – is defined as theĀ magnitude of a force which would cause a mass of one kg to accelerate with one meter per secondĀ per second. Per second per second. This is not a typo: 1 N corresponds to 1 kg times 1 m/sĀ per second, i.e. 1 kgĀ·m/s2. So… Because energy is force times distance, the unit of energyĀ may be expressed in units of kgĀ·m/s2Ā·m, or kgĀ·m2/s2, i.e. the unit of mass times the unit ofĀ velocity squared. To sum it all up:

1 J = 1 NĀ·m = 1 kgĀ·(m/s)2

This reflects the physical dimensionsĀ on both sides of theĀ E = mĀ·c2Ā formula again but… Well… How should weĀ interpretĀ this? Look at the animation below once more, and imagine the green dot is some tinyĀ massĀ moving around the origin, in an equally tiny circle. We’ve gotĀ twoĀ oscillations here: each packingĀ halfĀ of the total energy of… Well… Whatever it is that our elementary wavefunction might represent in realityĀ – which we don’t know, of course.

Now, the blue and the red dot – i.e. the horizontal and vertical projectionĀ of the green dot –Ā accelerate up and down. If we look carefully, we see these dots accelerateĀ towardsĀ the zero point and, once they’ve crossed it, theyĀ decelerate, so as to allow for a reversal of direction: the blue dot goes up, and then down. Likewise, the red dot does the same. The interplay between the two oscillations, because of the 90Ā° phase difference, is interesting: if the blue dot is at maximum speed (near or at the origin), the red dot reverses speed (its speed is, therefore, (almost) nil), and vice versa. The metaphor of our frictionless V-2 engine, our perpetuum mobile,Ā comes to mind once more.

The question is: what’s going on, really?

My answer is: I don’t know. I do think that, somehow, energy should be thought of as some two-dimensional oscillation of something – something which we refer to asĀ mass, but we didn’t define mass very clearly either. It also, somehow, combines linear and rotational motion. Each of the two dimensions packs half of the energy of the particle that is being represented by our wavefunction. It is, therefore, only logical that the physical unitĀ of both is to be expressed as a force over some distance – which is, effectively, the physical dimension of energy – or the rotational equivalent of them: torqueĀ over some angle.Ā Indeed, the analogy between linear and angular movement is obvious: theĀ kineticĀ energy of a rotating object is equal to K.E. = (1/2)Ā·IĀ·Ļ2. In this formula, I is the rotational inertiaĀ – i.e. the rotational equivalent of mass – and Ļ is the angular velocity – i.e. the rotational equivalent of linearĀ velocity. Noting that the (average) kinetic energy in any system must be equal to the (average) potential energy in the system, we can add both, so we get a formula which is structurallyĀ similar to theĀ E = mĀ·c2Ā formula. But isĀ it the same? Is the effective mass of some object the sum of an almost infinite number of quantaĀ that incorporate some kind ofĀ rotationalĀ motion? And – if we use the right units – is the angular velocity of these infinitesimally small rotations effectively equal to the speed of light?

I am not sure. Not at all, really. But, so far, I can’t think of any explanation of the wavefunction that would make more sense than this one. I just need to keep trying to find better ways toĀ articulateĀ orĀ imagineĀ what might be going on. š In this regard, I’d like to add a point – which may or may not be relevant. When I talked about that guitar string, or the water wave, and wrote that each point on the string – or each water drop – just moves up and down, we should think of the physicality of the situation: when the string oscillates, itsĀ lengthĀ increases. So it’s only because our string is flexible that it can vibrate between the fixed points at its ends. For a rope that’sĀ notĀ flexible, the end points would need to move in and out with the oscillation. Look at the illustration below, for example: the two kids who are holding rope must come closer to each other, so as to provide the necessary space inside of the oscillation for the other kid. šThe next illustration – of how water waves actually propagate – is, perhaps, more relevant. Just think of a two-dimensional equivalent – and of the two oscillations as being transverseĀ waves, as opposed to longitudinal.Ā See how string theory starts making sense? š

The most fundamental question remains the same: what is it,Ā exactly, that is oscillating here? What is theĀ field? It’s always some force on some charge – but what charge, exactly? Mass? What is it? Well… I don’t have the answer to that. It’s the same as asking: what isĀ electricĀ charge,Ā really? So the question is: what’s theĀ realityĀ of mass, of electric charge, or whatever other charge that causes a force toĀ actĀ on it?

If youĀ know, please letĀ meĀ know. š

Post scriptum: The fact that we’re talking someĀ two-dimensional oscillation here – think of a surface now – explains the probability formula: we need toĀ squareĀ the absolute value of the amplitude to get it. And normalize, of course. Also note that, when normalizing, we’d expect to get some factor involvingĀ Ļ somewhere, because we’re talking someĀ circularĀ surface – as opposed to a rectangular one. But I’ll letĀ youĀ figure that out. š

# Re-visiting electron orbitals (III)

In my previous post, I mentioned that it wasĀ not so obvious (both from a physicalĀ as well as from aĀ mathematicalĀ point of view) to write the wavefunction for electron orbitals – which we denoted as Ļ(x, t), i.e. a function of two variables (or four: one time coordinate and three space coordinates) –Ā as the product of two other functions in one variable only.

[…] OK. The above sentence is difficult to read. Let me write in math. š It isĀ notĀ so obvious to write Ļ(x, t) as:

Ļ(x, t) = eāiĀ·(E/Ä§)Ā·tĀ·Ļ(x)

As I mentioned before, the physicists’ use of the same symbol (Ļ, psi) for both the Ļ(x, t) and Ļ(x) function is quite confusing – because the two functions areĀ veryĀ different:

• Ļ(x, t) is a complex-valued function of twoĀ (real)Ā variables: x and t. OrĀ four, I should say, because xĀ = (x, y, z) – but it’s probably easier to think of x as oneĀ vectorĀ variable – aĀ vector-valued argument, so to speak. And then t is, of course, just aĀ scalarĀ variable. So… Well… A function of twoĀ variables: the position in space (x), and time (t).
• In contrast, Ļ(x) is a real-valuedĀ function ofĀ oneĀ (vector) variable only: x, so that’s the position in space only.

Now you should cry foul, of course: Ļ(x) is notĀ necessarilyĀ real-valued. It mayĀ be complex-valued. You’re right.Ā You know the formula:Note the derivation of this formula involved a switch from Cartesian to polar coordinates here, so from xĀ = (x, y, z) to rĀ = (r, Īø, Ļ), and that the function is also a function of the twoĀ quantum numbersĀ l and m now, i.e. the orbital angular momentum (l) and its z-component (m) respectively. In my previous post(s), I gave you the formulas for Yl,m(Īø, Ļ) and Fl,m(r) respectively. Fl,m(r) was a real-valued function alright, but the Yl,m(Īø, Ļ) had that eiĀ·mĀ·ĻĀ factor in it. So… Yes. You’re right: the Yl,m(Īø, Ļ) function is real-valued if – and onlyĀ if – m = 0, in which case eiĀ·mĀ·ĻĀ = 1.Ā Let me copy the table from Feynman’s treatment of the topic once again:The Plm(cosĪø) functions are the so-called (associated) Legendre polynomials, and the formula for these functions is rather horrible:Don’t worry about it too much: just note the Plm(cosĪø)Ā is aĀ real-valuedĀ function. The point is the following:theĀ Ļ(x, t) is a complex-valuedĀ function because – andĀ onlyĀ because – we multiply a real-valued envelope function – which depends on positionĀ only – with eāiĀ·(E/Ä§)Ā·tĀ·eiĀ·mĀ·ĻĀ = eāiĀ·[(E/Ä§)Ā·tĀ āĀ mĀ·Ļ].

[…]

Please read the above once again and – more importantly – think about it for a while. š You’ll have to agree with the following:

• As mentioned in my previous post,Ā the eiĀ·mĀ·ĻĀ factor just gives us phase shift: just aĀ re-set of our zero point for measuring time, so to speak, and the whole eāiĀ·[(E/Ä§)Ā·tĀ āĀ mĀ·Ļ]Ā factor just disappears when weāre calculating probabilities.
• The envelope function gives us the basic amplitude – in theĀ classicalĀ sense of the word:Ā the maximum displacement fromĀ theĀ zeroĀ value. And so it’s that eāiĀ·[(E/Ä§)Ā·tĀ āĀ mĀ·Ļ]Ā that ensures the whole expression somehow captures the energyĀ of the oscillation.

Let’s first look at the envelope function again. Let me copy the illustration forĀ n = 5 and lĀ = 2 from aĀ Wikimedia CommonsĀ article.Ā Note the symmetry planes:

• Any plane containing theĀ z-axis is a symmetry plane – like a mirror in which we can reflect one half of theĀ shape to get the other half. [Note that I am talking theĀ shapeĀ only here. Forget about the colors for a while – as these reflect the complex phase of the wavefunction.]
• Likewise, the plane containingĀ bothĀ the x– and the y-axis is a symmetry plane as well.

The first symmetry plane – or symmetryĀ line, really (i.e. theĀ z-axis) – should not surprise us, because the azimuthal angle Ļ is conspicuously absent in the formula for our envelope function if, as we are doing in this article here, we merge theĀ eiĀ·mĀ·ĻĀ factor with the eāiĀ·(E/Ä§)Ā·t, so it’s just part and parcel of what the author of the illustrations above refers to as the ‘complex phase’ of our wavefunction.Ā OK. Clear enough – I hope. š But why is theĀ the xy-plane a symmetry plane too? We need to look at that monstrous formula for the Plm(cosĪø) function here: just note the cosĪø argument in it is being squaredĀ before it’s used in all of the other manipulation. Now, we know that cosĪø = sin(Ļ/2Ā āĀ Īø). So we can define someĀ newĀ angle – let’s just call it Ī± – which is measured in the way we’re used to measuring angle, which is notĀ from the z-axis but from the xy-plane. So we write: cosĪø = sin(Ļ/2Ā āĀ Īø) = sinĪ±. The illustration below may or may not help you to see what we’re doing here.So… To make a long story short, we can substitute the cosĪø argument in the Plm(cosĪø) function for sinĪ± = sin(Ļ/2Ā āĀ Īø). Now, if the xy-plane is a symmetry plane, then we must find the same value for Plm(sinĪ±) and Plm[sin(āĪ±)]. Now, that’s not obvious, because sin(āĪ±) = āsinĪ± ā Ā sinĪ±. However, because the argument in that Plm(x) function is being squared before any other operation (like subtracting 1 and exponentiating the result), it is OK: [āsinĪ±]2Ā = [sinĪ±]2Ā =Ā sin2Ī±. […] OK, I am sure the geeks amongst my readers will be able to explain this more rigorously. In fact, I hope they’ll have a look at it, because there’s also that dl+m/dxl+mĀ operator, and so you should check what happens with the minus sign there. š

[…] Well… By now, you’re probably totally lost, but the fact of the matter is that we’ve got a beautiful result here. Let me highlight the most significant results:

• AĀ definiteĀ energy state of a hydrogen atom (or of an electron orbiting around some nucleus, I should say) appears to us as some beautifully shaped orbital – an envelopeĀ function in three dimensions, really – whichĀ has the z-axis – i.e. the vertical axis – as a symmetry line and the xy-plane as a symmetry plane.
• The eāiĀ·[(E/Ä§)Ā·tĀ āĀ mĀ·Ļ]Ā factor gives us the oscillation within the envelope function. As such, it’s this factor that, somehow,Ā captures the energyĀ of the oscillation.

It’s worth thinking about this. Look at the geometry of the situation again – as depicted below. We’re looking at the situation along the x-axis, in the direction of the origin, which is the nucleus of our atom.

The eiĀ·mĀ·ĻĀ factor just gives us phase shift: just aĀ re-set of our zero point for measuring time, so to speak. Interesting, weird – but probably less relevant than the eāiĀ·[(E/Ä§)Ā·tĀ factor, which gives us the two-dimensional oscillation that captures the energy of the state.

Now, the obvious question is: the oscillation of what, exactly? I am not quite sure but – as I explained in my Deep BlueĀ page – the real and imaginary part of our wavefunction are really like the electric and magnetic field vector of an oscillating electromagnetic field (think of electromagnetic radiation – if that makes it easier). Hence, just like the electric and magnetic field vector represent some rapidly changing forceĀ on a unit charge, the real and imaginary part of our wavefunction must also represent some rapidly changingĀ forceĀ on… Well… I am not quite sure on what though. The unit charge is usually defined as the charge of a proton – rather than an electron – but then forces act on some mass, right? And the massĀ of a proton is hugely different from the mass of an electron. The same electric (or magnetic) force will, therefore, give a hugely different acceleration to both.

So… Well… My guts instinct tells me the real and imaginary part of our wavefunction just represent, somehow, a rapidly changing force on some unit ofĀ mass, but then I am not sure how to define that unit right now (it’s probably notĀ the kilogram!).

Now, there is another thing we should note here: we’re actually sort of de-constructing a rotationĀ (look at the illustration above once again) in two linearly oscillating vectors – one along the z-axis and the other along the y-axis.Ā Hence, in essence, we’re actually talking about something that’s spinning.Ā In other words, we’re actually talking someĀ torqueĀ around the x-axis. In what direction? I think that shouldn’t matter – that we can write E or āE, in other words, but… Well… I need to explore this further – as should you! š

Let me just add one more note on the eiĀ·mĀ·ĻĀ factor. It sort of defines the geometryĀ of the complex phase itself. Look at the illustration below. Click on it to enlarge it if necessary – or, better still, visit the magnificent Wikimedia Commons article from which I get these illustrations. These are the orbitals nĀ = 4 and lĀ = 3. Look at the red hues in particular – or the blue – whatever: focus on one color only, and see how how – for mĀ = Ā±1, we’ve got one appearance of that color only. For mĀ = Ā±1, the same color appears at two ends of the ‘tubes’ – or toriĀ (plural of torus), I should say – just to sound more professional. š For mĀ = Ā±2, the torus consists of three parts – or, in mathematical terms, we’d say the order of its rotational symmetryĀ is equal to 3.Ā Check that Wikimedia Commons article for higher values ofĀ nĀ andĀ l: the shapes become very convoluted, but the observation holds. š

Have fun thinking all of this through for yourself – and please do look at those symmetries in particular. š

Post scriptum: You should do some thinking on whether or not theseĀ mĀ =Ā Ā±1, Ā±2,…, Ā±lĀ orbitals are really different. As I mentioned above, a phase difference is just what it is: a re-set of the t = 0 point. Nothing more, nothing less. So… Well… As far as I am concerned, that’s notĀ aĀ realĀ difference, is it? š As with other stuff, I’ll let you think about this for yourself.

# Feynman’s Seminar on Superconductivity (1)

The ultimate challenge for students of Feynman’s iconic LecturesĀ series is, of course, to understand his final one: A Seminar on Superconductivity. As he notes in his introduction to this formidably dense piece, the text does notĀ present the detail of each and every step in the development and, therefore,Ā we’re not supposed to immediately understand everything. As Feynman puts it: we should just believe (more or less) that things would come out if we wouldĀ be able to go through each and every step. Well… Let’s see. Feynman throws a lot of stuff in hereāincluding, I suspect, some stuff that may not beĀ directlyĀ relevant, but that he sort of couldn’t insert into all of his otherĀ Lectures. So where do we start?

It took me one long maddening day to figure out the first formula:It says that the amplitude for a particle to go from aĀ toĀ bĀ in a vector potential (think of a classical magnetic field) is the amplitude for the same particle to go fromĀ aĀ toĀ b when there is no field (A = 0) multiplied by the exponential of the line integral of the vector potential times the electric charge divided by Planck’s constant. I stared at this for quite a while, but thenĀ I recognized the formula for the magnetic effect on an amplitude, which I described in my previous post, which tells us that a magnetic field will shiftĀ the phase of the amplitude of a particle with an amount equal to:

Hence, if we write ā©b|aāŖ for A = 0 as ā©b|aāŖA = 0Ā = CĀ·eiĪø, then ā©b|aāŖĀ in AĀ will, naturally, be equal to ā©b|aāŖĀ in A = CĀ·ei(Īø+Ļ)Ā = CĀ·eiĪøĀ·eiĻĀ =Ā ā©b|aāŖA = 0Ā Ā·eiĻ, and so that explains it. š Alright… Next. Or… Well… Let us briefly re-examine the concept of the vector potential, because we’ll need it a lot. We introduced it in our post on magnetostatics. Let’s briefly re-cap the development there. In Maxwell’s set of equations, two out of the four equations give us the magnetic field:Ā āā¢BĀ = 0 andĀ c2āĆBĀ =Ā j/Īµ0. We noted the following in this regard:

1. TheĀ āā¢BĀ = 0 equation is true,Ā always, unlike theĀ āĆEĀ = 0 expression, which is true for electrostatics only (no moving charges). So theĀ āā¢BĀ = 0 equation says theĀ divergenceĀ ofĀ BĀ is zero,Ā always.
2. The divergence of the curl of a vector field is always zero. Hence, if AĀ is some vector field, thenĀ div(curlĀ A) =Ā āā¢(āĆA) = 0,Ā always.
3. We can now apply another theorem: if the divergence of a vector field, sayĀ D, is zeroāso ifĀ āā¢DĀ = 0āthenĀ DĀ will be theĀ the curl of some other vector fieldĀ C, so we can write:Ā DĀ =Ā āĆC. Ā Applying this toĀ āā¢BĀ = 0, we can write:Ā

IfĀ āā¢BĀ = 0, then there is anĀ AĀ such thatĀ BĀ =Ā āĆA

So, in essence, we’re just re-defining the magnetic field (B) in terms of some other vector field. To be precise, we write it as the curlĀ of some other vector field, which we refer to asĀ the (magnetic)Ā vectorĀ potential. The components of the magnetic field vector can then be re-written as:

We need to note an important point here: the equations above suggest that the components of BĀ depend on position only. In other words, we assumeĀ staticĀ magnetic fields, so theyĀ doĀ notĀ change with time. That, in turn, assumes steady currents. We will want to extend the analysis to also includeĀ magnetodynamics. It complicates the analysis but… Well… Quantum mechanicsĀ isĀ complicated. Let us remind ourselves here of Feynman’s re-formulation of Maxwell’s equations as a set ofĀ twoĀ equations (expressed in terms of the magnetic (vector) and the electric potential) only:

These equations are wave equations, as you can see by writing out the second equation:

It is a wave equation in three dimensions. Note that, even in regions where we do no have any charges or currents, we have non-zeroĀ solutions for ĻĀ andĀ A. These non-zero solutions are, effectively, representing the electric and magnetic fields as they travel through free space. As Feynman notes, the advantage of re-writing Maxwell’s equations as we do above, is that the two new equations make it immediately apparent that we’re talking electromagnetic waves, really. As he notes, for many practical purposes, it will still be convenient to use the original equations in terms of E and B, but… Well… Not in quantum mechanics, it turns out. As Feynman puts it: “E and BĀ are on the other side of the mountain we have climbed. Now we are ready to cross over to the other side of the peak. Things will look differentāwe are ready for some new and beautiful views.”

Well… Maybe. Appreciating those views, as part of our study of quantum mechanics, does take time and effort, unfortunately. š¦

### The SchrĆ¶dinger equation in an electromagnetic field

Feynman then jots down SchrĆ¶dinger’s equation for the same particle (with chargeĀ q) moving in an electromagnetic field that is characterized not only by the (scalar) potentialĀ Ī¦ but also by a vector potentialĀ A:

Now where doesĀ thatĀ come from? We know the standard formula in anĀ electricĀ field, right? It’s the formula we used to find the energy states of electrons in a hydrogen atom:

iĀ·Ä§Ā·āĻ/āt = ā(1/2)Ā·(Ä§2/m)ā2Ļ + VĀ·Ļ

Of course, it is easy to see that we replaced V by qĀ·Ī¦, which makes sense: the potential of a charge in an electric field is the product of the charge (q) and the (electric) potential (Ī¦), because Ī¦ is, obviously, the potential energy of the unitĀ charge. It’s also easy to see we can re-write āÄ§2Ā·ā2Ļ as [(Ä§/i)Ā·ā]Ā·[(Ä§/i)Ā·ā]Ļ because (1/i)Ā·(1/i) = 1/i2Ā = 1/(ā1) = ā1. š Alright. So it’s just that āqĀ·AĀ term in the (Ä§/i)ā ā qĀ·AĀ expression that we need to explain now.

Unfortunately, that explanation isĀ notĀ so easy. Feynman basically re-derives SchrĆ¶dinger’s equation using his trade-markĀ historical argument – which did notĀ include any magnetic field –Ā withĀ a vector potential. The re-derivation is rather annoying, and I didn’t have the courage to go through it myself, so you should – just like me – just believe Feynman when he says that, when there’s a vector potential – i.e. when there’s a magnetic field – then that (Ä§/i)Ā·ā operatorĀ – which is the momentum operator– ought to be replaced by a new momentum operator:

So… Well… There we are… š So far, so good? Well… Maybe.

While, as mentioned, you won’t be interested in the mathematical argument, it is probably worthwhile to reproduce Feynman’s more intuitive explanation of why the operator above is what it is. In other words, let us try to understand thatĀ āqA term. Look at the following situation: we’ve got a solenoid here, and some current I is going through it so there’s a magnetic field B. Think of theĀ dynamicsĀ while weĀ turn onĀ this flux. Maxwell’s second equation (āĆE =Ā āāB/āt) tells us theĀ line integral ofĀ EĀ around a loop will be equal to the timeĀ rate of change of the magnetic flux through that loop. TheĀ āĆE =Ā āāB/āt equation is aĀ differentialĀ equation, of course, so it doesn’t have the integral, but you get the ideaāI hope.

Now, using the BĀ =Ā āĆA equationĀ we can re-write theĀ āĆE =Ā āāB/āt asĀ āĆE =Ā āā(āĆA)/āt. This allows us to write the following:

Ā āĆE =Ā āā(āĆA)/āt =Ā āāĆ(āA/āt)Ā ā E =Ā āāA/āt

This is a remarkable expression. Note its derivation is based on the commutativity of the curl and time derivative operators, which is a property that can easily be explained: if we have a function in two variablesāsay x and tāthen the order of the derivation doesn’t matter: we can first take the derivative with respect toĀ xĀ and then toĀ tĀ or, alternatively, we can first take the time derivative and then do theĀ ā/āx operation. So… Well… The curl is, effectively, a derivative with regard to the spatial variables. OK. So what? What’s the point?

Well… If we’d have some charge q, as shown in the illustration above, that would happen to be there as the flux is being switched on, it will experience a force which is equal to F = qE. We can now integrate this over the time interval (t) during which the flux is being built up to get the following:

ā«0tĀ F =Ā ā«0tĀ mĀ·a =Ā ā«0tĀ mĀ·dv/dt =Ā mĀ·vt= ā«0tĀ qĀ·EĀ =Ā āā«0tĀ qĀ·āA/āt =Ā āqĀ·At

AssumingĀ v0Ā and A0Ā are zero, we may drop the time subscript and simply write:

mĀ·vĀ =Ā āqĀ·A

The point is: during the build-up of the magnetic flux, our charge will pick up some (classical)Ā momentumĀ that is equal to p =Ā mĀ·vĀ =Ā āqĀ·A. So… Well… That sort of explains the additional term in our new momentum operator.

Note: For some reason I don’t quite understand, Feynman introduces the weird concept of ‘dynamical momentum’, which he defines as the quantityĀ mĀ·vĀ + qĀ·A, so that quantity must be zero in the analysis above. I quickly googled to see why but didn’t invest too much time in the research here. It’s just… Well… A bit puzzling. I don’t really see the relevance of his point here: I am quite happy to go along with the new operator, as it’s rather obvious that introducingĀ changingĀ magnetic fields must, obviously, also have some impact on our wave equationsāin classical as well as in quantum mechanics.

### Local conservation of probability

The title of this section in Feynman’s Lecture (yes, still the same Lecture – we’re not switching topics here) is theĀ equation of continuity for probabilities. I find it brilliant, because it confirmsĀ myĀ interpretation of the wave function as describing some kind of energy flow. Let me quote Feynman on his endeavor here:

“An important part of the SchrĆ¶dinger equation for a single particle is the idea that the probability to find the particle at a position is given by the absolute square of the wave function. It is also characteristic of the quantum mechanics that probability is conserved in a local sense. When the probability of finding the electron somewhere decreases, while the probability of the electron being elsewhere increases (keeping the total probability unchanged), something must be going on in between. In other words, the electron has a continuity in the sense that if the probability decreases at one place and builds up at another place, there must be some kind of flow between. If you put a wall, for example, in the way, it will have an influence and the probabilities will not be the same. So the conservation of probability alone is not the complete statement of the conservation law, just as the conservation of energy alone is not as deep and important as the local conservation of energy.Ā If energy is disappearing, there must be a flow of energy to correspond. In the same way, we would like to find a ācurrentā of probability such that if there is any change in the probability density (the probability of being found in a unit volume), it can be considered as coming from an inflow or an outflow due to some current.”

This is it, really ! The wave function does represent some kind of energy flow – between a so-called ‘real’ and a so-called ‘imaginary’ space, which are to be defined in terms of directional versus rotational energy, as I try to point out – admittedly: more by appealing to intuition than to mathematical rigor – in that post of mine on the meaning of the wavefunction.

So what is theĀ flowĀ – or probabilityĀ currentĀ as Feynman refers to it? Well… Here’s the formula:

Huh?Ā Yes. Don’t worry too much about it right now. The essential point is to understand what this current – denoted by J – actually stands for:

So what’s next? Well… Nothing. I’ll actually refer you to Feynman now, because I can’t improve on howĀ heĀ explains howĀ pairs of electronsĀ start behaving when temperatures are low enough to render Boltzmann’s Law irrelevant: the kinetic energy that’s associated with temperatureĀ can no longer break up electron pairs if temperature comes close to the zero point.

Huh? What? Electron pairs? Electrons are not supposed to form pairs, are they? They carry the same charge and are, therefore, supposed to repel each other. Well… Yes and no. In my post on the electron orbitals in a hydrogen atom – which just presented Feynman’s presentationĀ on the subject-matter in a, hopefully, somewhat more readable format – we calculated electron orbitals neglecting spin. In Feynman’s words:

“We make anotherĀ approximation by forgetting that the electron has spin. […] The non-relativistic SchrĆ¶dinger equation disregards magnetic effects. [However] Small magnetic effects [do] occur because, from the electronās point-of-view, the proton is a circulating charge which produces a magnetic field. In this field the electron will have a different energy with its spin up than with it down. [Hence] The energy of the atom will be shifted a little bit from what we will calculate. We will ignore this small energy shift. Also we will imagine that the electron is just like a gyroscope moving around in space always keeping the same direction of spin. Since we will be considering a free atom in space the total angular momentum will be conserved. In our approximation we will assume that the angular momentum of the electron spin stays constant, so all the rest of the angular momentum of the atomāwhat is usually called āorbitalā angular momentumāwill also be conserved. To an excellent approximation the electron moves in the hydrogen atom like a particle without spināthe angular momentum of the motion is a constant.”

To an excellent approximation… But… Well… Electrons in a metal do form pairs, because they can give up energy in that way and, hence, they are more stableĀ that way. Feynman does not go into the details here – I guess because that’s way beyond the undergrad level – but refers to the Bardeen-Coopers-Schrieffer (BCS) theory instead – the authors of which got a Nobel Prize in Physics in 1972 (that’s a decade or so after Feynman wrote this particularĀ Lecture), so I must assume the theory is well accepted now. š

Of course, you’ll shout now: Hey! Hydrogen is not a metal!Ā Well… Think again: the latest breakthrough in physics is making hydrogen behave like a metal. š And I am really talking the latestĀ breakthrough: ScienceĀ just published the findings of this experiment last month! š š In any case, we’re not talking hydrogen here but superconducting materials, to which – as far as we know – the BCS theory does apply.

So… Well… I am done. I just wanted to show you why it’s important to work your way through Feynman’s last LectureĀ because… Well… Quantum mechanics does explain everything – although the nitty-gritty of it (the Meissner effect, the London equation, flux quantization, etc.) are rather hard bullets to bite. š¦

Don’t give up ! I am struggling with the nitty-gritty too ! š

# An interpretation of the wavefunction

This is my umpteenth post on the same topic. š¦ It is obvious that this search for a sensibleĀ interpretation is consuming me. Why? I am not sure. Studying physics is frustrating. As a leading physicist puts it:

“TheĀ teaching of quantum mechanics these days usuallyĀ follows the same dogma: firstly, the student is told about the failure of classical physics atĀ the beginning of the last century; secondly, the heroic confusions of the founding fathersĀ are described and the student is given to understand that no humble undergraduate studentĀ could hope to actually understand quantum mechanics for himself; thirdly, a deus exĀ machina arrives in the form of a set of postulates (the SchrĆ¶dinger equation, the collapseĀ of the wavefunction, etc); fourthly, a bombardment of experimental verifications is given,Ā so that the student cannot doubt that QM is correct; fifthly, the student learns how toĀ solve the problems that will appear on the exam paper, hopefully with as little thought asĀ possible.”

That’s obviously not the way we want to understand quantum mechanics. [With we,Ā I mean, me, of course, and you, if you’re reading this blog.]Ā Of course, that doesn’t mean I don’t believe Richard Feynman, one of the greatest physicists ever, when he tells us no one, including himself, understands physics quite the way we’dĀ likeĀ to understand it. Such statements should not prevent us from tryingĀ harder. So let’s look for betterĀ metaphors.Ā The animation below shows the two components of the archetypal wavefunction ā a simple sine and cosine. They’re the same function actually, but their phases differ by 90 degrees (Ļ/2).

It makes me think of a V-2 engine with the pistons at a 90-degree angle. Look at the illustration below, which I took from a rather simple article on cars and engines that has nothing to do with quantum mechanics. Think of the moving pistons as harmonic oscillators, like springs.

We will also think of theĀ center of each cylinder as the zero point: think of that point as a point where – if we’re looking at one cylinder alone – the internal and external pressure balance each other, so the piston would not move… Well… If it weren’t for the other piston, because the second piston isĀ not at the centerĀ when the first is. In fact, it is easy to verify and compare the following positions of both pistons, as well as the associated dynamics of the situation:

 Piston 1 Piston 2 Motion of Piston 1 Motion Piston 2 Top Center Compressed air will push piston down Piston moves down against external pressure Center Bottom Piston moves down against external pressure External air pressure will push piston up Bottom Center External air pressure will push piston up Piston moves further up and compresses the air Center Top Piston moves further up and compresses the air Compressed air will push piston down

When the pistons move, their linear motion will be described by a sinusoidal function: a sine or a cosine. In fact, the 90-degree V-2 configuration ensures that the linear motion of the two pistons will be exactly the same, except for a phase difference of 90 degrees. [Of course, because of the sideways motion of the connecting rods, our sine and cosine function describes the linear motion only approximately, but you can easily imagine the idealizedĀ limit situation. If not, check Feynman’s description of the harmonic oscillator.]

The question is: if we’d have a set-up like this, two springs – or two harmonic oscillators – attached to a shaft through a crank, would this really work as a perpetuum mobile? We obviously talk energyĀ being transferred back and forth between the rotating shaft and the moving pistons… So… Well… Let’s model this: the totalĀ energy, potentialĀ andĀ kinetic, in each harmonic oscillator is constant. Hence, the piston only delivers or receivesĀ kineticĀ energy from the rotating mass of the shaft.

Now, in physics, that’s a bit of an oxymoron: we don’t think of negative or positive kinetic (or potential) energy in the context of oscillators. We don’t think of the direction of energy. But… Well… If we’ve got twoĀ oscillators, our picture changes, and so we may have to adjust our thinking here.

Let me start by giving you an authoritative derivation of the various formulas involved here, taking the example of the physical spring as an oscillatorābut the formulas are basically the same forĀ any harmonic oscillator.

The first formula is a general description of the motion of our oscillator. The coefficient in front of the cosine function (a)Ā is the maximum amplitude. Of course, you will also recognize Ļ0Ā as theĀ naturalĀ frequency of the oscillator, andĀ Ī as the phase factor, which takes into account our t = 0 point. In our case, for example, we have two oscillators with a phase difference equal to Ļ/2 and, hence, Ī would be 0 for one oscillator, and āĻ/2 for the other. [The formula to apply here is sinĪø = cos(Īø ā Ļ/2).] Also note that we can equate our Īø argument to Ļ0Ā·t.Ā Now, ifĀ aĀ = 1 (which is the case here), then these formulas simplify to:

1. K.E. = T = mĀ·v2/2 =Ā mĀ·Ļ02Ā·sin2(Īø + Ī) = mĀ·Ļ02Ā·sin2(Ļ0Ā·t + Ī)
2. P.E. = U = kĀ·x2/2 = kĀ·cos2(Īø + Ī)

The coefficient k in the potential energy formula characterizes the force: F = ākĀ·x. The minus sign reminds us our oscillator wants to return to the center point, so the force pulls back. From the dynamics involved, it is obvious that k must be equal to mĀ·Ļ02., so that gives us the famous T + U = mĀ·Ļ02/2 formula or, including aĀ once again, T + U = mĀ·a2Ā·Ļ02/2.

Now, if we normalizeĀ our functions by equating k to one (k = 1), thenĀ the motion ofĀ our first oscillator is given by the cosĪø function, and its kinetic energy will be equal toĀ sin2Īø. Hence, the (instantaneous)Ā changeĀ in kinetic energy at any point in time will be equal to:

d(sin2Īø)/dĪø = 2āsinĪøād(sinĪø)/dt = 2āsinĪøācosĪø

Let’s look at the second oscillator now. Just think of the second piston going up and down in our V-twin engine. Its motion is given by theĀ sinĪø function which, as mentioned above, is equal to cos(ĪøāĻ /2). Hence, its kinetic energy is equal toĀ sin2(ĪøāĻ /2), and how itĀ changesĀ – as a function of Īø – will be equal to:

2āsin(ĪøāĻ /2)ācos(ĪøāĻ /2) =Ā = ā2ācosĪøāsinĪø = ā2āsinĪøācosĪø

We have our perpetuum mobile! While transferring kinetic energy from one piston to the other, the rotating shaft moves at constant speed. Linear motion becomes circular motion, and vice versa, in a frictionless Universe. We have the metaphor we were looking for!

Somehow, in this beautiful interplay between linear and circular motion, energy is being borrowed from one place to another, and then returned. From what place to what place? I am not sure. We may call it the real and imaginary energy space respectively, but what does that mean? One thing is for sure, however: the interplay between the real and imaginary part of the wavefunction describes how energy propagates through space!

How exactly? Again, I am not sure. Energy is, obviously, mass in motion – as evidenced by the E = mĀ·c2Ā equation, and it may not have any direction (when everything is said and done, it’s a scalarĀ quantity without direction), but the energy in a linear motion is surely different from that in a circular motion, and our metaphor suggests we need to think somewhat more along those lines. Perhaps we will, one day, able toĀ square this circle. š

SchrĆ¶dinger’s equation

Let’s analyze the interplay between the real and imaginary part of the wavefunction through an analysis of SchrĆ¶dinger’s equation, which we write as:

iĀ·Ä§āāĻ/āt = ā(Ä§2/2m)āā2Ļ + VĀ·Ļ

We can do a quick dimensional analysis of both sides:

• [iĀ·Ä§āāĻ/āt] = Nāmās/s = Nām
• [ā(Ä§2/2m)āā2Ļ] = Nām3/m2 = Nām
• [VĀ·Ļ] = Nām

Note the dimension of the ‘diffusion’ constantĀ Ä§2/2m: [Ä§2/2m] = N2ām2ās2/kg = N2ām2ās2/(NĀ·s2/m) = Nām3. Also note that, in order for the dimensions to come out alright, the dimension of V ā the potential ā must be that of energy. Hence, Feynmanās description of it as the potential energy ā rather than the potential tout courtĀ ā is somewhat confusing but correct: V must equal the potential energy of the electron. Hence, V is not the conventional (potential) energy of the unit charge (1 coulomb). Instead, the natural unit of charge is used here, i.e. the charge of the electron itself.

Now, SchrĆ¶dingerās equation ā without the VĀ·Ļ term ā can be written as the following pair of equations:

1. Re(āĻ/āt) = ā(1/2)ā(Ä§/m)āIm(ā2Ļ)
2. Im(āĻ/āt) = (1/2)ā(Ä§/m)āRe(ā2Ļ)

This closely resembles the propagation mechanism of an electromagnetic wave as described by Maxwell’s equation for free space (i.e. a space with no charges), but E and B are vectors, not scalars. How do we get this result. Well… Ļ is a complex function, which we can write as a + iāb. Likewise, āĻ/āt is a complex function, which we can write as c + iād, and ā2Ļ can then be written as e + iāf. If we temporarily forget about the coefficients (Ä§, Ä§2/m and V), then SchrĆ¶dingerās equation – including VĀ·Ļ term – amounts to writing something like this:

iā(c + iād) = ā(e + iāf) + (a + iāb) ā a + iāb = iāc ā d + e+ iāf Ā ā a = ād + e and b = c + f

Hence, we can now write:

1. VāRe(Ļ) = āÄ§āIm(āĻ/āt) + (1/2)ā( Ä§2/m)āRe(ā2Ļ)
2. VāIm(Ļ) = Ä§āRe(āĻ/āt) + (1/2)ā( Ä§2/m)āIm(ā2Ļ)

This simplifies to the two equations above for V = 0, i.e. when there is no potential (electron in free space). Now we can bring the Re and Im operators into the brackets to get:

1. VāRe(Ļ) = āÄ§āāIm (Ļ)/āt + (1/2)ā( Ä§2/m)āā2Re(Ļ)
2. VāIm(Ļ) = Ä§āāRe(Ļ)/āt + (1/2)ā( Ä§2/m)āā2Im(Ļ)

This is very interesting, because we can re-write this using the quantum-mechanical energy operator H = ā(Ä§2/2m)āā2 + VĀ· (note the multiplication sign after the V, which we do not have ā for obvious reasons ā for the ā(Ä§2/2m)āā2 expression):

1. H[Re (Ļ)] = āÄ§āāIm(Ļ)/āt
2. H[Im(Ļ)] = Ä§āāRe(Ļ)/āt

A dimensional analysis shows us both sides are, once again, expressed in Nām. Itās a beautiful expression because ā if we write the real and imaginary part of Ļ as rācosĪø and rāsinĪø, we get:

1. H[cosĪø] = āÄ§āāsinĪø/āt = EācosĪø
2. H[sinĪø] = Ä§āācosĪø/āt = EāsinĪø

Indeed, ĪøĀ = (Eāt ā pāx)/Ä§ and, hence, āÄ§āāsinĪø/āt = Ä§ācosĪøāE/Ä§ = EācosĪø and Ä§āācosĪø/āt = Ä§āsinĪøāE/Ä§ = EāsinĪø.Ā  Now we can combine the two equations in one equation again and write:

H[rā(cosĪø + iāsinĪø)] = rā(EācosĪø + iāsinĪø) ā H[Ļ] = EāĻ

The operator H ā applied to the wavefunction ā gives us the (scalar) product of the energy E and the wavefunction itself. Isn’t this strange?

Hmm… I need to further verify and explain this result… I’ll probably do so in yet another post on the same topic… š

Post scriptum: The symmetry of our V-2 engine – or perpetuum mobileĀ – is interesting: its cross-section has only one axis of symmetry. Hence, we may associate some angle with it, so as to define its orientation in the two-dimensional cross-sectional plane. Of course, the cross-sectional plane itself is at right angles to the crankshaft axis, which we may also associate with some angle in three-dimensional space. Hence, its geometry defines two orthogonal directions which, in turn, define a spherical coordinate system, as shown below.

We may, therefore, say that three-dimensional space is actually being impliedĀ byĀ the geometry of our V-2 engine. Now thatĀ isĀ interesting, isn’t it? š

# All what you ever wanted to know about the photon wavefunction…

Post scriptum note added on 11 July 2016: This is one of the more speculative posts which led to my e-publication analyzing the wavefunction as an energy propagation. With the benefit of hindsight, I would recommend you to immediately read the more recentĀ exposĆ©Ā on the matter that is being presented here, which you can find by clicking on the provided link.

Original post:

This post is, essentially, a continuation of my previous post, in which I juxtaposed the following images:

Both are the same, and then theyāre not. The illustration on the right-hand side is a regular quantum-mechanical wavefunction, i.e. an amplitudeĀ wavefunction. You’ve seen that one before. In this case, the x-axis represents time, so we’re looking at the wavefunction at some particular point in space. ]You know we can just switch the dimensions and it would all look the same.] The illustration on the left-hand side looks similar, but it’s notĀ an amplitude wavefunction. The animationĀ shows how the electric field vector (E) of an electromagnetic wave travels through space. Its shape is the same. So it’s the same function. Is it also the same reality?

Yes and no. And I would say: more no than yesāin this case, at least. Note that the animation doesĀ notĀ show the accompanying magnetic field vector (B). That vector is equally essential in the electromagnetic propagation mechanism according to Maxwellās equations, whichālet me remind youāare equal to:

1. āB/āt = āāĆE
2. āE/āt = āĆB

In fact, I should writeĀ the second equation as āE/āt = c2āĆB, but then I assume we measure time and distance in equivalent units, so c = 1.

You know that E and B are two aspects of one and the same thing: if we have one, then we have the other. To be precise, B is always orthogonal to EĀ in the direction thatās given by the right-hand rule for the following vector cross-product: B = exĆE,Ā with exĀ the unit vector pointing in the x-direction (i.e. the direction of propagation). The reality behind is illustrated below for aĀ linearly polarized electromagnetic wave.

TheĀ B = exĆEĀ equation is equivalent to writing B= iĀ·E, which is equivalent to:

BĀ =Ā iĀ·E = ei(Ļ/2)Ā·ei(kx ā Ļt)Ā = cos(kx ā Ļt + Ļ/2) +Ā iĀ·sin(kx ā Ļt + Ļ/2)

=Ā āsin((kx ā Ļt) + iĀ·cos(kx ā Ļt)

Now,Ā E and B have only two components: EyĀ andĀ Ez, and ByĀ and Bz. That’s only because we’re looking at some idealĀ orĀ elementaryĀ electromagnetic wave here but… Well… Let’s just go along with it. š It is then easy to prove that the equation above amounts to writing:

1. ByĀ = cos(kx ā Ļt + Ļ/2) = āsin(kx ā Ļt) = āEz
2. BzĀ = sin(kx ā Ļt + Ļ/2) = cos(kx ā Ļt) = Ey

We should now think of EyĀ andĀ EzĀ as the real and imaginary part of some wavefunction, which we’ll denote as ĻEĀ = ei(kx ā Ļt). So we write:

E = (Ey,Ā Ez) =Ā EyĀ + iĀ·EzĀ = cos(kx ā Ļt) + iāsin(kx ā Ļt) =Ā Re(ĻE) +Ā iĀ·Im(ĻE) = ĻEĀ = ei(kx ā Ļt)

What about B? We just do the same, so we write:

BĀ = (By, Bz) = ByĀ + iĀ·BzĀ = ĻBĀ = iĀ·E = iĀ·ĻEĀ = āsin(kx ā Ļt) + iāsin(kx ā Ļt) = āĀ Im(ĻE) +Ā iĀ·Re(ĻE)

Now weĀ need to prove that ĻEĀ andĀ ĻBĀ are regular wavefunctions, which amounts to proving SchrĆ¶dingerās equation, i.e.Ā āĻ/āt =Ā iĀ·(Ä§/m)Ā·ā2Ļ, forĀ both ĻEĀ andĀ ĻB. [Note I use the SchrĆ¶dingerās equation for a zero-mass spin-zero particle here, which uses the Ä§/m factor rather than the Ä§/(2m) factor.] To prove that ĻEĀ andĀ ĻBĀ are regular wavefunctions, we should prove that:

1. Re(āĻE/āt) = Ā ā(Ä§/m)Ā·Im(ā2ĻE) andĀ Im(āĻE/āt) = (Ä§/m)Ā·Re(ā2ĻE), and
2. Re(āĻB/āt) = Ā ā(Ä§/m)Ā·Im(ā2ĻB) andĀ Im(āĻB/āt) = (Ä§/m)Ā·Re(ā2ĻB).

Letās do the calculations for the second pair of equations. TheĀ time derivative on the left-hand side is equal to:

āĻB/āt = āiĻĀ·iei(kx ā Ļt)Ā =Ā ĻĀ·[cos(kx ā Ļt) + iĀ·sin(kx ā Ļt)] = ĻĀ·cos(kx ā Ļt) +Ā iĻĀ·sin(kx ā Ļt)

The second-order derivative on the right-hand side is equal to:

ā2ĻBĀ =Ā ā2ĻB/āx2Ā = iĀ·k2Ā·ei(kx ā Ļt)Ā = k2Ā·cos(kx ā Ļt) + iĀ·k2Ā·sin(kx ā Ļt)

So the two equations forĀ ĻBĀ are equivalent to writing:

1. Re(āĻB/āt) = Ā  ā(Ä§/m)Ā·Im(ā2ĻB)Ā āĀ ĻĀ·cos(kx ā Ļt) =Ā k2Ā·(Ä§/m)Ā·cos(kx ā Ļt)
2. Im(āĻB/āt) = (Ä§/m)Ā·Re(ā2ĻB)Ā ā ĻĀ·sin(kx ā Ļt) = k2Ā·(Ä§/m)Ā·sin(kx ā Ļt)

So we see that both conditions are fulfilled if, and only if, Ļ = k2Ā·(Ä§/m).

Now, we also demonstrated in that post of mineĀ that Maxwellās equations imply the following:

1. āBy/āt =Ā ā(āĆE)yĀ = āEz/āx =Ā ā[sin(kx ā Ļt)]/āx = kĀ·cos(kx ā Ļt) = kĀ·Ey
2. āBz/āt =Ā ā(āĆE)zĀ =Ā ā āEy/āx =Ā ā ā[cos(kx ā Ļt)]/āx =Ā kĀ·sin(kx ā Ļt) = kĀ·Ez

Hence, using those ByĀ = āEzĀ andĀ BzĀ = EyĀ equations above, we can also calculate these derivatives as:

1. āBy/āt = āāEz/āt =Ā āāsin(kx ā Ļt)/āt = ĻĀ·cos(kx ā Ļt) = ĻĀ·Ey
2. āBz/āt = āEy/āt = ācos(kx ā Ļt)/āt = āĻĀ·[āsin(kx ā Ļt)] = ĻĀ·Ez

In other words, Maxwellās equations imply that Ļ = k, which is consistent with us measuring time and distance in equivalent units, so the phase velocity is Ā cĀ = 1 =Ā Ļ/k.

So far, so good. We basically established that the propagation mechanism for an electromagnetic wave, as described by Maxwellās equations, is fully coherent with the propagation mechanismāif we can call it like thatāas described by SchrĆ¶dingerās equation. We also established the following equalities:

1. Ļ = k
2. Ļ = k2Ā·(Ä§/m)

The second of the twoĀ de Broglie equations tells us that k = p/Ä§, so we can combineĀ these two equations and re-write these two conditions as:

Ļ/k = 1Ā = kĀ·(Ä§/m) = (p/Ä§)Ā·(Ä§/m) = p/mĀ ā p = m

What does this imply? The p here is the momentum: p = mĀ·v, so this condition implies vĀ must be equal to 1 too, so the wave velocity is equal to the speed of light. Makes sense, because we actuallyĀ areĀ talking light here. š In addition, because it’s light, we also know E/p =Ā cĀ = 1, so we have – once again – the general E = p = m equation, which we’ll need!

OK. Next. Let’s write the SchrĆ¶dinger wave equation for both wavefunctions:

1. āĻE/āt =Ā iĀ·(Ä§/mE)Ā·ā2ĻE, and
2. āĻB/āt =Ā iĀ·(Ä§/mB)Ā·ā2ĻB.

Huh?Ā What’sĀ mEĀ andĀ mE? We should only associate one mass concept with our electromagnetic wave, shouldn’t we? Perhaps. I just want to be on the safe side now. Of course, if we distinguishĀ mEĀ and mB, we should probably also distinguish pEĀ and pB, and EEĀ and EBĀ as well, right? Well… Yes. If we accept this line of reasoning, then the mass factor in SchrĆ¶dinger’s equations is pretty much like the 1/c2Ā = Ī¼0Īµ0Ā factor in Maxwellās (1/c2)Ā·āE/āt = āĆBĀ equation: the mass factor appears as a property of the medium, i.e. theĀ vacuumĀ here! [Just check my post on physical constants in case you wonder what I am trying to say here, in which I explain why and howĀ cĀ definesĀ the (properties of the) vacuum.]

To be consistent, we should also distinguish pEĀ and pB, and EEĀ and EB, and so we should write ĻEĀ and ĻBĀ as:

1. ĻEĀ = ei(kEx ā ĻEt), and
2. ĻBĀ = ei(kBx ā ĻBt).

Huh?Ā Yes.Ā I know what you think: weāre talking one photonāor one electromagnetic waveāso there can be only one energy, one momentum and, hence, only one k, and one Ļ. Wellā¦ Yes and no. Of course, the following identities should hold: kEĀ = kBĀ and, likewise, ĻEĀ =Ā ĻB. Soā¦ Yes. Theyāre the same: one k and one Ļ. But thenā¦ Wellā¦ Conceptually, the two kās and Ļās are different. So we write:

1. pEĀ = EEĀ = mE, and
2. pBĀ = EBĀ = mB.

The obvious question is: can we just add them up to find the totalĀ energy and momentum of our photon? The answer is obviously positive: E = EEĀ + EB, pĀ = pEĀ + pBĀ and mĀ = mEĀ + mB.

Letās check a few things now. How does it work for the phase and group velocity of ĻEĀ and ĻB? Simple:

1. vg = āĻE/ākEĀ = ā[EE/Ä§]/ā[pE/Ä§] = āEE/āpE = āpE/āpEĀ = 1
2. vp = ĻE/kE =Ā (EE/Ä§)/(pE/Ä§) = EE/pE = pE/pE = 1

So weāre fine, and you can check the result for ĻBĀ by substituting the subscript E for B. To sum it all up, what weāve got here is the following:

1. We can think of a photon having some energy thatās equal to E = p = mĀ (assuming c = 1), but that energy would be split up in an electric and a magnetic wavefunction respectively:Ā ĻEĀ and ĻB.
2. SchrĆ¶dingerās equation applies to bothĀ wavefunctions, but the E, p and m in those two wavefunctions are the same and not the same: their numericalĀ value is the same (pEĀ =EEĀ = mEĀ = pBĀ =EBĀ = mB), but theyāre conceptuallyĀ different. They must be: if not, weād get a phase and group velocity for the wave that doesnāt make sense.

Of course, the phase and group velocity for theĀ sumĀ of the ĻEĀ and ĻBĀ waves must also be equal toĀ c. This is obviously the case, because weāre adding waves with the same phase and group velocity c, so thereās no issue with the dispersion relation.

So let’sĀ insert those pEĀ =EEĀ = mEĀ = pBĀ =EBĀ = mBĀ values in the two wavefunctions. For ĻE, we get:

ĻEĀ = ei[kEx ā ĻEt)Ā =Ā ei[(pE/Ä§)Ā·x ā (EE/Ä§)Ā·t]Ā

You can do the calculation for ĻBĀ yourself. Let’s simplify our life a little bit and assume we’re using Planck units, so Ä§ = 1, and so the wavefunction simplifies to ĻEĀ =Ā eiĀ·(pEĀ·x ā EEĀ·t). We can now add the components of E and BĀ using the summation formulas for sines and cosines:

1. ByĀ + EyĀ = cos(pBĀ·xĀ āĀ EBĀ·t + Ļ/2)Ā + cos(pEĀ·xĀ āĀ EEĀ·t) = 2Ā·cos[(pĀ·x ā EĀ·t + Ļ/2)/2]Ā·cos(Ļ/4) = ā2Ā·cos(pĀ·x/2 ā EĀ·t/2 + Ļ/4)

2. BzĀ + EzĀ = sin(pBĀ·xĀ āĀ EBĀ·t+Ļ/2) + sin(pEĀ·xĀ āĀ EEĀ·t) =Ā 2Ā·sin[(pĀ·x ā EĀ·t + Ļ/2)/2]Ā·cos(Ļ/4) =Ā ā2Ā·sin(pĀ·x/2 ā EĀ·t/2 + Ļ/4)

Interesting!Ā We find aĀ compositeĀ wavefunction for our photon which we can write as:

E + B =Ā ĻEĀ + ĻBĀ = EĀ +Ā iĀ·EĀ =Ā ā2Ā·ei(pĀ·x/2 ā EĀ·t/2 + Ļ/4)Ā =Ā ā2Ā·ei(Ļ/4)Ā·ei(pĀ·x/2 ā EĀ·t/2)Ā =Ā ā2Ā·ei(Ļ/4)Ā·E

What a great result! Itās easy to double-check, because we can see the EĀ +Ā iĀ·EĀ = ā2Ā·ei(Ļ/4)Ā·EĀ formula implies that 1 +Ā iĀ should equalĀ ā2Ā·ei(Ļ/4). Now thatās easy to prove, both geometrically (just do a drawing) or formally: ā2Ā·ei(Ļ/4)Ā = ā2Ā·cos(Ļ/4) + iĀ·sin(Ļ/4ei(Ļ/4)Ā = (ā2/ā2) + iĀ·(ā2/ā2) = 1 + i. WeāreĀ bang on!Ā š

We can double-check once more, because we should get the same from addingĀ E and BĀ = iĀ·E, right? Letās try:

EĀ +Ā BĀ =Ā EĀ +Ā iĀ·E = cos(pEĀ·xĀ āĀ EEĀ·t) + iĀ·sin(pEĀ·xĀ āĀ EEĀ·t) +Ā iĀ·cos(pEĀ·xĀ āĀ EEĀ·t) āĀ sin(pEĀ·xĀ āĀ EEĀ·t)

= [cos(pEĀ·xĀ āĀ EEĀ·t) ā sin(pEĀ·xĀ āĀ EEĀ·t)] +Ā iĀ·[sin(pEĀ·xĀ āĀ EEĀ·t) ā cos(pEĀ·xĀ āĀ EEĀ·t)]

Indeed, we can see weāre going to obtain the same result, because the āsinĪø in the real part of our compositeĀ wavefunction is equal toĀ cos(Īø+Ļ/2), and the ācosĪø in its imaginary part is equal to sin(Īø+Ļ/2). So the sum above is the same sum of cosines and sines that we did already.

So our electromagnetic wavefunction, i.e. the wavefunction for theĀ photon, is equal to:

ĻĀ =Ā ĻEĀ + ĻBĀ = ā2Ā·ei(pĀ·x/2 ā EĀ·t/2 + Ļ/4)Ā = ā2Ā·ei(Ļ/4)Ā·ei(pĀ·x/2 ā EĀ·t/2)Ā

What about the ā2 factor in front, and theĀ Ļ/4 term in the argument itself? No sure. It must have something to do with the way the magnetic force works, which is notĀ like the electric force. Indeed, remember the Lorentz formula: the force on some unit charge (q = 1) will be equal to F = E + vĆB. Soā¦ Wellā¦ Weāve got another cross-product here and so the geometry of the situation is quite complicated: itāsĀ notĀ like adding two forces F1Ā andĀ F2Ā to get some combined force F = F1Ā andĀ F2.

In any case, we need the energy, and we know that its proportional to the square of the amplitude, soā¦ Wellā¦ Weāre spot on: the square of the ā2 factor in the ā2Ā·cos product and ā2Ā·sin product is 2, so thatās twiceā¦ Wellā¦ What? Hold on a minute!Ā Weāre actually taking theĀ absoluteĀ square of the E + B =Ā ĻEĀ + ĻBĀ = EĀ +Ā iĀ·EĀ =Ā ā2Ā·ei(pĀ·x/2 ā EĀ·t/2 + Ļ/4)Ā wavefunction here. Is thatĀ legal? I must assume it isāalthoughā¦ Wellā¦ Yes. You’re right. We should do some more explaining here.

We knowĀ that we usually measure the energy as someĀ definiteĀ integral, from t = 0 to some other point in time, or over the cycle of the oscillation. So whatās theĀ cycleĀ here? Our combined wavefunction can be written asĀ ā2Ā·ei(pĀ·x/2 ā EĀ·t/2 + Ļ/4)Ā = ā2Ā·ei(Īø/2 + Ļ/4), so a full cycle would correspond to Īø going from 0 to 4Ļ here, rather than from 0 to 2Ļ. So that explains the ā2 factor in front of our wave equation.

Bingo! If you were looking for an interpretation of the Planck energy and momentum, here it is. And, while everything thatās written above is not easy to understand, itās close to the āintuitiveā understanding to quantum mechanics that we were looking for, isnāt it? The quantum-mechanical propagation model explainsĀ everything now. šĀ I only need to show one more thing, and thatās the different behaviorĀ of bosons and fermions:

1. The amplitudes of identiticalĀ bosonic particles interfere with a positive sign, so we have Bose-Einstein statistics here. As Feynman writes it: (amplitude direct) + (amplitude exchanged).
2. The amplitudes of identicalĀ fermionicĀ particles interfere with a negative sign, so we have Fermi-Dirac statistics here: (amplitude direct) ā (amplitude exchanged).

Iāll think about it. I am sure itās got something to do with that B= iĀ·EĀ formula or, to put it simply, with the fact that, when bosons are involved, we get two wavefunctions (ĻEĀ and ĻB) for the price of one. The reasoning should be something like this:

I. For a massless particle (i.e. a zero-massĀ fermion), our wavefunction is just Ļ =Ā ei(pĀ·x ā EĀ·t). So we have no ā2 or ā2Ā·ei(Ļ/4)Ā factor in front here. So we can just add any number of them āĀ Ļ1Ā + Ļ2Ā + Ļ3Ā + ā¦ ā and then take the absolute square of the amplitude to find a probability density, and weāre done.

II. For a photon (i.e. a zero-massĀ boson), our wavefunction is ā2Ā·ei(Ļ/4)Ā·ei(pĀ·x ā EĀ·t)/2, which ā letās introduce a new symbol ā weāll denote by Ļ, so Ļ = ā2Ā·ei(Ļ/4)Ā·ei(pĀ·x ā EĀ·t)/2. Now, if we add any number of these, we get a similar sum but with that ā2Ā·ei(Ļ/4)Ā factor in front, so we write: Ļ1Ā + Ļ2Ā + Ļ3Ā + ā¦ = ā2Ā·ei(Ļ/4)Ā·(Ļ1Ā + Ļ2Ā + Ļ3Ā + ā¦). If we take the absolute square now, weāll see the probability density will be equal to twiceĀ the density for the Ļ1Ā + Ļ2Ā + Ļ3Ā + ā¦ sum, because

|ā2Ā·ei(Ļ/4)Ā·(Ļ1Ā + Ļ2Ā + Ļ3Ā + ā¦)|2Ā = |ā2Ā·ei(Ļ/4)|2Ā·|Ļ1Ā + Ļ2Ā + Ļ3Ā + ā¦)|2Ā =Ā 2Ā·|Ļ1Ā + Ļ2Ā + Ļ3Ā + ā¦)|2

Soā¦ Wellā¦ I still need to connect this to Feynmanās (amplitude direct) Ā± (amplitude exchanged) formula, but I am sure it can be done.

Now, we haven’t tested the completeĀ ā2Ā·ei(Ļ/4)Ā·ei(pĀ·x ā EĀ·t)/2Ā wavefunction. Does it respect SchrĆ¶dinger’s āĻ/āt = iĀ·(1/m)Ā·ā2Ļ or, including the 1/2 factor, theĀ āĻ/āt = iĀ·[1/2m)]Ā·ā2Ļ equation? [Note we assume, once again, that Ä§ = 1, so we use Planck units once more.] Let’s see. We can calculate the derivatives as:

• āĻ/āt =Ā āā2Ā·ei(Ļ/4)Ā·eāiā[pĀ·x ā EĀ·t]/2Ā·(iĀ·E/2)
• ā2Ļ =Ā ā2[ā2Ā·ei(Ļ/4)Ā·eāiā[pĀ·x ā EĀ·t]/2]/āx2Ā = ā2Ā·ei(Ļ/4)Ā·ā[ā2Ā·ei(Ļ/4)Ā·eāiā[pĀ·x ā EĀ·t]/2Ā·(iĀ·p/2)]/āx = āā2Ā·ei(Ļ/4)Ā·eāiā[pĀ·x ā EĀ·t]/2Ā·(p2/4)

So SchrĆ¶dinger’s equation becomes:

āiĀ·ā2Ā·ei(Ļ/4)Ā·eāiā[pĀ·x ā EĀ·t]/2Ā·(iĀ·E/2) = āiĀ·(1/m)Ā·ā2Ā·ei(Ļ/4)Ā·eāiā[pĀ·x ā EĀ·t]/2Ā·(p2/4) ā 1/2 = 1/4!?

That’s funny ! It doesn’t work ! The E and m and p2Ā are OK because we’ve got that E = m = p equation, but we’ve got problems with yet another factor 2. It only works when we use the 2/m coefficient in SchrĆ¶dinger’s equation.

So… Well… There’s no choice. That’s what we’re going to do. TheĀ SchrĆ¶dinger equation for the photon is āĻ/āt = iĀ·(2/m)Ā·ā2Ļ !

It’s a very subtle point. This is all great, and veryĀ fundamental stuff! Let’s now move on to SchrĆ¶dinger’s actualĀ equation, i.e. theĀ āĻ/āt = iĀ·(Ä§/2m)Ā·ā2Ļ equation.

Post scriptum on the Planck units:

If we measure time and distance in equivalent units, say seconds, we can re-write the quantum of action as:

1.0545718Ć10ā34Ā NĀ·mĀ·s = (1.21Ć1044Ā N)Ā·(1.6162Ć10ā35Ā m)Ā·(5.391Ć10ā44Ā s)

ā (1.0545718Ć10ā34/2.998Ć108) NĀ·s2Ā = (1.21Ć1044Ā N)Ā·(1.6162Ć10ā35/2.998Ć108Ā s)(5.391Ć10ā44Ā s)

ā (1.21Ć1044Ā N) = [(1.0545718Ć10ā34/2.998Ć108)]/[(1.6162Ć10ā35/2.998Ć108Ā s)(5.391Ć10ā44Ā s)]Ā NĀ·s2/s2

You’ll say: what’s this? Well… Look at it. We’ve got a much easier formula for the Planck forceāmuch easier than the standard formulas you’ll find on Wikipedia, for example. If we re-interpret the symbolsĀ Ä§ andĀ cĀ so they denote the numericalĀ value of the quantum of action and the speed of light in standard SI units (i.e. newton, meter and second)āso Ä§ andĀ c become dimensionless, orĀ mathematicalĀ constants only, rather thanĀ physicalĀ constantsāthen the formula above can be written as:

FPĀ newtonĀ = (Ä§/c)/[(lP/c)Ā·tP] newtonĀ ā FPĀ = Ä§/(lPĀ·tP)

Just double-check it: 1.0545718Ć10ā34/(1.6162Ć10ā35Ā·5.391Ć10ā44) = 1.21Ć1044. Bingo!

You’ll say: what’s the point? The point is: our model is complete. We don’t need the other physical constants ā i.e. the Coulomb, Boltzmann and gravitational constant āĀ to calculate the Planck units we need, i.e. the Planck force, distance and time units. It all comes out of our elementary wavefunction! All we need to explain the UniverseĀ ā or, let’s be more modest, quantum mechanicsĀ ā is two numerical constants (c and Ä§) and Euler’s formula (which uses Ļ andĀ e, of course). That’s it.

If you don’t think that’s a great result, then… Well… Then you’re not reading this. š