One Equation, Too Many Jobs: Rethinking Schrödinger’s Equation and Wavefunction

I have just republished one of my long-standing papers on de Broglie’s matter-wave concept as a new, standalone publication, with its own DOI:

👉 De Broglie’s matter-wave concept and issues
https://www.researchgate.net/publication/399225854_De_Broglie’s_matter-wave_concept_and_issues
DOI: 10.13140/RG.2.2.30104.25605

The reason for republishing is not cosmetic. A new Annex was added on 31 December 2025 that fundamentally clarified — for me, at least — what Schrödinger’s equation is really doing, and just as importantly, what it is not doing.

This clarification came out of a long and at times uncomfortable dialogue with the most recent version of OpenAI’s GPT model (ChatGPT 5.2). Uncomfortable, because it initially destabilized a view I had held for years. Productive, because it forced a deeper structural distinction that I now believe is unavoidable. Let me explain.


The uncomfortable admission: I was wrong about the 12\tfrac{1}{2} factor​

For a long time, I was convinced that the factor 12\tfrac{1}{2} factor in Schrödinger’s equation — especially in the hydrogen atom problem — must reflect some deeper pairing mechanism. At times, I even wondered whether the equation was implicitly modeling an electron pair (opposite spin), rather than a single electron.

That intuition was not random. It came from a broader realist programme in which I treat the electron as a structured object, with internal dynamics (zitterbewegung-like orbital motion), not as a point particle. If mass, energy, and phase all have internal structure, why should a simple quadratic kinetic term with a mysterious 12\tfrac{1}{2} be fundamental?

The hard truth is this: that intuition was misplaced — but it was pointing in the right direction.

The mistake was not questioning the factor 12\tfrac{1}{2}. The mistake was assuming Schrödinger’s equation was trying to describe everything at once.


The key insight: Schrödinger describes the envelope, not the engine

The decisive realization was structural:

Schrödinger’s wavefunction does not describe the electron’s internal dynamics.
It describes the translational envelope of phase coherence.

Once you see that, several things fall into place immediately:

  • The hydrogen “orbitals” are not literal orbits, and not internal electron motion.
  • They are standing-wave solutions of an envelope phase, constrained by a Coulomb potential.
  • The factor 12\tfrac{1}{2}​ is not mysterious at all at this level: it is the natural coefficient that appears in effective, averaged, quadratic envelope dynamics.

In other words:
The 12\tfrac{1}{2} factor belongs to the envelope layer, not to the internal structure of the electron.

My earlier “electron pair” idea tried to explain a structural feature by inventing new ontology. The correct move was simpler and more radical: separate the layers.


One symbol, too many jobs

Modern quantum mechanics makes a profound — and in my view costly — simplification:

It uses one symbol, ψ, to represent:

  • internal phase,
  • translational dynamics,
  • probability amplitudes,
  • and experimental observables.

That compression works operationally, but it hides structure.

What the new Annex makes explicit is that Nature almost certainly does not work that way. At minimum, we should distinguish:

  1. Internal phase
    Real, physical, associated with internal orbital motion and energy bookkeeping.
  2. Envelope phase
    Slow modulation across space, responsible for interference, diffraction, and spectra.
  3. Observables
    What experiments actually measure, which are sensitive mainly to envelope-level phase differences.

Once this distinction is made, long-standing confusions dissolve rather than multiply.


Why this does not contradict experiments

This is crucial.

Nothing in this reinterpretation invalidates:

  • electron diffraction,
  • hydrogen spectra,
  • interference experiments,
  • or the empirical success of standard quantum mechanics.

On the contrary: it explains why Schrödinger’s equation works so well — within its proper domain.

The equation is not wrong.
It is just over-interpreted.


A personal note on changing one’s mind

I’ll be honest: this line of reasoning initially felt destabilizing. It challenged a position I had defended for years. But that discomfort turned out to be a feature, not a bug.

Good theory-building does not preserve intuitions at all costs. It preserves structure, coherence, and explanatory power.

What emerged is a cleaner picture:

  • internal realism without metaphysics,
  • Schrödinger demoted from “ultimate truth” to “effective envelope theory”,
  • and a much clearer map of where different mathematical tools belong.

That, to me, is progress.


Where this opens doors

Once we accept that one wavefunction cannot represent all layers of Nature, new possibilities open up:

  • clearer interpretations of spin and the Dirac equation,
  • better realist models of lattice propagation,
  • a more honest treatment of “quantum mysteries” as category mistakes,
  • and perhaps new mathematical frameworks that respect internal structure from the start.

Those are not promises — just directions.

For now, I am satisfied that one long-standing conceptual knot has been untied.

And sometimes, that’s enough for a good year’s work. 🙂

Post Scriptum: On AI, Intellectual Sparring, and the Corridor

A final remark, somewhat orthogonal to physics.

The revision that led to this blog post and the accompanying paper did not emerge from a sudden insight, nor from a decisive experimental argument. It emerged from a long, occasionally uncomfortable dialogue with an AI system, in which neither side “won,” but both were forced to refine their assumptions.

At the start of that dialogue, the AI responded in a largely orthodox way, reproducing standard explanations for the factor 12\tfrac{1}{2}​ in Schrödinger’s equation. I, in turn, defended a long-held intuition that this factor must point to internal structure or pairing. What followed was not persuasion, but sparring: resistance on both sides, followed by a gradual clarification of conceptual layers. The breakthrough came when it became clear that a single mathematical object — the wavefunction — was being asked to do too many jobs at once.

From that moment on, the conversation shifted from “who is right?” to “which layer are we talking about?” The result was not a victory for orthodoxy or for realism, but a structural separation: internal phase versus translational envelope, engine versus modulation. That separation resolved a tension that had existed for years in my own thinking.

I have explored this mode of human–AI interaction more systematically in a separate booklet on ResearchGate, where I describe such exchanges as occurring within a corridor: a space in which disagreement does not collapse into dominance or deference, but instead forces both sides toward finer distinctions and more mature reasoning.

This episode convinced me that the real intellectual value of AI does not lie in answers, but in sustained resistance without ego — and in the willingness of the human interlocutor to tolerate temporary destabilization without retreating into dogma. When that corridor holds, something genuinely new can emerge.

In that sense, this post is not only about Schrödinger’s equation. It is also about how thinking itself may evolve when humans and machines are allowed to reason together, rather than merely agree.

Readers interested in this kind of human–AI interaction beyond the present physics discussion may want to look at that separate booklet I published on ResearchGate (≈100 pages), in which I try to categorize different modes of AI–human intellectual interaction — from superficial compliance and authority projection to genuine sparring. In that text, exchanges like the one briefly alluded to above are described as a Type-D collapse: a situation in which both human and AI are forced to abandon premature explanatory closure, without either side “winning,” and where progress comes from structural re-layering rather than persuasion.

The booklet is intentionally exploratory and occasionally playful in tone, but it grew out of exactly this kind of experience: moments where resistance, rather than agreement, turns out to be the most productive form of collaboration.

Feynman’s Seminar on Superconductivity (1)

Pre-script (dated 26 June 2020): This post got mutilated by the removal of some material by the dark force. You should be able to follow the main story line, however. If anything, the lack of illustrations might actually help you to think things through for yourself. In any case, we now have different views on these concepts as part of our realist interpretation of quantum mechanics, so we recommend you read our recent papers instead of these old blog posts.

Original post:

The ultimate challenge for students of Feynman’s iconic Lectures series is, of course, to understand his final one: A Seminar on Superconductivity. As he notes in his introduction to this formidably dense piece, the text does not present the detail of each and every step in the development and, therefore, we’re not supposed to immediately understand everything. As Feynman puts it: we should just believe (more or less) that things would come out if we would be able to go through each and every step. Well… Let’s see. Feynman throws a lot of stuff in here—including, I suspect, some stuff that may not be directly relevant, but that he sort of couldn’t insert into all of his other Lectures. So where do we start?

It took me one long maddening day to figure out the first formula:f1It says that the amplitude for a particle to go from to in a vector potential (think of a classical magnetic field) is the amplitude for the same particle to go from to b when there is no field (A = 0) multiplied by the exponential of the line integral of the vector potential times the electric charge divided by Planck’s constant. I stared at this for quite a while, but then I recognized the formula for the magnetic effect on an amplitude, which I described in my previous post, which tells us that a magnetic field will shift the phase of the amplitude of a particle with an amount equal to:

integral

Hence, if we write 〈b|a〉 for A = 0 as 〈b|aA = 0 = C·eiθ, then 〈b|a〉 in A will, naturally, be equal to 〈b|a〉 in A = C·ei(θ+φ) = C·eiθ·eiφ = 〈b|aA = 0 ·eiφ, and so that explains it. 🙂 Alright… Next. Or… Well… Let us briefly re-examine the concept of the vector potential, because we’ll need it a lot. We introduced it in our post on magnetostatics. Let’s briefly re-cap the development there. In Maxwell’s set of equations, two out of the four equations give us the magnetic field: B = 0 and c2×B = j0. We noted the following in this regard:

  1. The ∇B = 0 equation is true, always, unlike the ×E = 0 expression, which is true for electrostatics only (no moving charges). So the B = 0 equation says the divergence of B is zero, always.
  2. The divergence of the curl of a vector field is always zero. Hence, if A is some vector field, then div(curl A) = •(×A) = 0, always.
  3. We can now apply another theorem: if the divergence of a vector field, say D, is zero—so if D = 0—then will be the the curl of some other vector field C, so we can write: D = ×C.  Applying this to B = 0, we can write: 

If B = 0, then there is an A such that B = ×A

So, in essence, we’re just re-defining the magnetic field (B) in terms of some other vector field. To be precise, we write it as the curl of some other vector field, which we refer to as the (magnetic) vector potential. The components of the magnetic field vector can then be re-written as:

formula for B

We need to note an important point here: the equations above suggest that the components of B depend on position only. In other words, we assume static magnetic fields, so they do not change with time. That, in turn, assumes steady currents. We will want to extend the analysis to also include magnetodynamics. It complicates the analysis but… Well… Quantum mechanics is complicated. Let us remind ourselves here of Feynman’s re-formulation of Maxwell’s equations as a set of two equations (expressed in terms of the magnetic (vector) and the electric potential) only:

Wave equation for A

Wave equation for potential

These equations are wave equations, as you can see by writing out the second equation:

wave equation

It is a wave equation in three dimensions. Note that, even in regions where we do no have any charges or currents, we have non-zero solutions for φ and A. These non-zero solutions are, effectively, representing the electric and magnetic fields as they travel through free space. As Feynman notes, the advantage of re-writing Maxwell’s equations as we do above, is that the two new equations make it immediately apparent that we’re talking electromagnetic waves, really. As he notes, for many practical purposes, it will still be convenient to use the original equations in terms of E and B, but… Well… Not in quantum mechanics, it turns out. As Feynman puts it: “E and B are on the other side of the mountain we have climbed. Now we are ready to cross over to the other side of the peak. Things will look different—we are ready for some new and beautiful views.”

Well… Maybe. Appreciating those views, as part of our study of quantum mechanics, does take time and effort, unfortunately. 😦

The Schrödinger equation in an electromagnetic field

Feynman then jots down Schrödinger’s equation for the same particle (with charge q) moving in an electromagnetic field that is characterized not only by the (scalar) potential Φ but also by a vector potential A:

schrodinger

Now where does that come from? We know the standard formula in an electric field, right? It’s the formula we used to find the energy states of electrons in a hydrogen atom:

i·ħ·∂ψ/∂t = −(1/2)·(ħ2/m)∇2ψ + V·ψ

Of course, it is easy to see that we replaced V by q·Φ, which makes sense: the potential of a charge in an electric field is the product of the charge (q) and the (electric) potential (Φ), because Φ is, obviously, the potential energy of the unit charge. It’s also easy to see we can re-write −ħ2·∇2ψ as [(ħ/i)·∇]·[(ħ/i)·∇]ψ because (1/i)·(1/i) = 1/i2 = 1/(−1) = −1. 🙂 Alright. So it’s just that −q·A term in the (ħ/i)∇ − q·A expression that we need to explain now.

Unfortunately, that explanation is not so easy. Feynman basically re-derives Schrödinger’s equation using his trade-mark historical argument – which did not include any magnetic field – with a vector potential. The re-derivation is rather annoying, and I didn’t have the courage to go through it myself, so you should – just like me – just believe Feynman when he says that, when there’s a vector potential – i.e. when there’s a magnetic field – then that (ħ/i)·∇ operator – which is the momentum operator– ought to be replaced by a new momentum operator:

new-momentum-operator

So… Well… There we are… 🙂 So far, so good? Well… Maybe.

While, as mentioned, you won’t be interested in the mathematical argument, it is probably worthwhile to reproduce Feynman’s more intuitive explanation of why the operator above is what it is. In other words, let us try to understand that −qA term. Look at the following situation: we’ve got a solenoid here, and some current I is going through it so there’s a magnetic field B. Think of the dynamics while we turn on this flux. Maxwell’s second equation (∇×E = −∂B/∂t) tells us the line integral of E around a loop will be equal to the time rate of change of the magnetic flux through that loop. The ∇×E = −∂B/∂t equation is a differential equation, of course, so it doesn’t have the integral, but you get the idea—I hope.solenoid

Now, using the B = ×A equation we can re-write the ∇×E = −∂B/∂t as ∇×E = −∂(×A)/∂t. This allows us to write the following:

 ∇×E = −∂(×A)/∂t = −×(∂A/∂t) ⇔ E = −∂A/∂t

This is a remarkable expression. Note its derivation is based on the commutativity of the curl and time derivative operators, which is a property that can easily be explained: if we have a function in two variables—say x and t—then the order of the derivation doesn’t matter: we can first take the derivative with respect to and then to t or, alternatively, we can first take the time derivative and then do the ∂/∂x operation. So… Well… The curl is, effectively, a derivative with regard to the spatial variables. OK. So what? What’s the point?

Well… If we’d have some charge q, as shown in the illustration above, that would happen to be there as the flux is being switched on, it will experience a force which is equal to F = qE. We can now integrate this over the time interval (t) during which the flux is being built up to get the following:

0t F = ∫0t m·a = ∫0t m·dv/dt = m·vt= ∫0t q·E = −∫0t q·∂A/∂t = −q·At

Assuming v0 and Aare zero, we may drop the time subscript and simply write:

v = −q·A

The point is: during the build-up of the magnetic flux, our charge will pick up some (classical) momentum that is equal to p = m·v = −q·A. So… Well… That sort of explains the additional term in our new momentum operator.

Note: For some reason I don’t quite understand, Feynman introduces the weird concept of ‘dynamical momentum’, which he defines as the quantity m·v + q·A, so that quantity must be zero in the analysis above. I quickly googled to see why but didn’t invest too much time in the research here. It’s just… Well… A bit puzzling. I don’t really see the relevance of his point here: I am quite happy to go along with the new operator, as it’s rather obvious that introducing changing magnetic fields must, obviously, also have some impact on our wave equations—in classical as well as in quantum mechanics.

Local conservation of probability

The title of this section in Feynman’s Lecture (yes, still the same Lecture – we’re not switching topics here) is the equation of continuity for probabilities. I find it brilliant, because it confirms my interpretation of the wave function as describing some kind of energy flow. Let me quote Feynman on his endeavor here:

“An important part of the Schrödinger equation for a single particle is the idea that the probability to find the particle at a position is given by the absolute square of the wave function. It is also characteristic of the quantum mechanics that probability is conserved in a local sense. When the probability of finding the electron somewhere decreases, while the probability of the electron being elsewhere increases (keeping the total probability unchanged), something must be going on in between. In other words, the electron has a continuity in the sense that if the probability decreases at one place and builds up at another place, there must be some kind of flow between. If you put a wall, for example, in the way, it will have an influence and the probabilities will not be the same. So the conservation of probability alone is not the complete statement of the conservation law, just as the conservation of energy alone is not as deep and important as the local conservation of energy. If energy is disappearing, there must be a flow of energy to correspond. In the same way, we would like to find a “current” of probability such that if there is any change in the probability density (the probability of being found in a unit volume), it can be considered as coming from an inflow or an outflow due to some current.”

This is it, really ! The wave function does represent some kind of energy flow – between a so-called ‘real’ and a so-called ‘imaginary’ space, which are to be defined in terms of directional versus rotational energy, as I try to point out – admittedly: more by appealing to intuition than to mathematical rigor – in that post of mine on the meaning of the wavefunction.

So what is the flow – or probability current as Feynman refers to it? Well… Here’s the formula:

probability-current-2

Huh? Yes. Don’t worry too much about it right now. The essential point is to understand what this current – denoted by J – actually stands for:

probability-current-1

So what’s next? Well… Nothing. I’ll actually refer you to Feynman now, because I can’t improve on how he explains how pairs of electrons start behaving when temperatures are low enough to render Boltzmann’s Law irrelevant: the kinetic energy that’s associated with temperature can no longer break up electron pairs if temperature comes close to the zero point.

Huh? What? Electron pairs? Electrons are not supposed to form pairs, are they? They carry the same charge and are, therefore, supposed to repel each other. Well… Yes and no. In my post on the electron orbitals in a hydrogen atom – which just presented Feynman’s presentation on the subject-matter in a, hopefully, somewhat more readable format – we calculated electron orbitals neglecting spin. In Feynman’s words:

“We make another approximation by forgetting that the electron has spin. […] The non-relativistic Schrödinger equation disregards magnetic effects. [However] Small magnetic effects [do] occur because, from the electron’s point-of-view, the proton is a circulating charge which produces a magnetic field. In this field the electron will have a different energy with its spin up than with it down. [Hence] The energy of the atom will be shifted a little bit from what we will calculate. We will ignore this small energy shift. Also we will imagine that the electron is just like a gyroscope moving around in space always keeping the same direction of spin. Since we will be considering a free atom in space the total angular momentum will be conserved. In our approximation we will assume that the angular momentum of the electron spin stays constant, so all the rest of the angular momentum of the atom—what is usually called “orbital” angular momentum—will also be conserved. To an excellent approximation the electron moves in the hydrogen atom like a particle without spin—the angular momentum of the motion is a constant.”

To an excellent approximation… But… Well… Electrons in a metal do form pairs, because they can give up energy in that way and, hence, they are more stable that way. Feynman does not go into the details here – I guess because that’s way beyond the undergrad level – but refers to the Bardeen-Coopers-Schrieffer (BCS) theory instead – the authors of which got a Nobel Prize in Physics in 1972 (that’s a decade or so after Feynman wrote this particular Lecture), so I must assume the theory is well accepted now. 🙂

Of course, you’ll shout now: Hey! Hydrogen is not a metal! Well… Think again: the latest breakthrough in physics is making hydrogen behave like a metal. 🙂 And I am really talking the latest breakthrough: Science just published the findings of this experiment last month! 🙂 🙂 In any case, we’re not talking hydrogen here but superconducting materials, to which – as far as we know – the BCS theory does apply.

So… Well… I am done. I just wanted to show you why it’s important to work your way through Feynman’s last Lecture because… Well… Quantum mechanics does explain everything – although the nitty-gritty of it (the Meissner effect, the London equation, flux quantization, etc.) are rather hard bullets to bite. 😦

Don’t give up ! I am struggling with the nitty-gritty too ! 🙂

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/