Cleaning Up After Bell

On the limits of theorems, the sociology of prizes, and the slow work of intellectual maturity

When I re-read two older posts of mine on Bell’s Theorem — one written in 2020, at a moment when my blog was gaining unexpected traction, and another written in 2023 in reaction to what I then experienced as a Nobel Prize award controversy — I feel a genuine discomfort.

Not because I think the core arguments were wrong.
But because I now see more clearly what was doing the talking.

There is, in both texts, a mixture of three things:

  1. A principled epistemic stance (which is still there);
  2. A frustration with institutional dynamics in physics (also there);
  3. But, yes, also a degree of rhetorical impatience that no longer reflects how I want to think — or be read.

This short text is an attempt to disentangle those layers.


1. Why I instinctively refused to “engage” with Bell’s Theorem

In the 2020 post, I wrote — deliberately provocatively — that I “did not care” about Bell’s Theorem. That phrasing was not chosen to invite dialogue; it was chosen to draw a boundary. At the time, my instinctive reasoning was this:

Bell’s Theorem is a mathematical theorem. Like any theorem, it tells us what follows if certain premises are accepted. Its physical relevance therefore depends entirely on whether those premises are physically mandatory, or merely convenient formalizations.

This is not a rejection of mathematics. It is a refusal to grant mathematics automatic ontological authority.

I was — and still am — deeply skeptical of the move by which a formal result is elevated into a metaphysical verdict about reality itself. Bell’s inequalities constrain a particular class of models (local hidden-variable models of a specific type). They do not legislate what Nature must be. In that sense, my instinct was aligned not only with Einstein’s well-known impatience with axiomatic quantum mechanics, but also with Bell himself, who explicitly hoped that a “radical conceptual renewal” might one day dissolve the apparent dilemma his theorem formalized.

Where I now see a weakness is not in the stance, but in its expression. Saying “I don’t care” reads as dismissal, while what I really meant — and should have said — is this:

I do not accept the premises as ontologically compulsory, and therefore I do not treat the theorem as decisive.

That distinction matters.


2. Bell, the Nobel Prize, and a sociological paradox

My 2023 reaction was sharper, angrier, and less careful — and that is where my current discomfort is strongest.

At the time, it seemed paradoxical to me that:

  • Bell was once close to receiving a Nobel Prize for a theorem he himself regarded as provisional,
  • and that nearly six decades later, a Nobel Prize was awarded for experiments demonstrating violations of Bell inequalities.

In retrospect, the paradox is not logical — it is sociological.

The 2022 Nobel Prize did not “disprove Bell’s Theorem” in a mathematical sense. It confirmed, experimentally and with great technical sophistication, that Nature violates inequalities derived under specific assumptions. What was rewarded was experimental closure, not conceptual resolution.

The deeper issue — what the correlations mean — remains as unsettled as ever.

What troubled me (and still does) is that the Nobel system has a long history of rewarding what can be stabilized experimentally, while quietly postponing unresolved interpretational questions. This is not scandalous; it is structural. But it does shape the intellectual culture of physics in ways that deserve to be named.

Seen in that light, my indignation was less about Bell, and more about how foundational unease gets ritualized into “progress” without ever being metabolized conceptually.


3. Authority, responsibility, and where my anger really came from

The episode involving John Clauser and climate-change denial pushed me from critique into anger — and here, too, clarity comes from separation.

The problem there is not quantum foundations.
It is the misuse of epistemic authority across domains.

A Nobel Prize in physics does not confer expertise in climate science. When prestige is used to undermine well-established empirical knowledge in an unrelated field, that is not dissent — it is category error dressed up as courage.

My reaction was visceral because it touched a deeper nerve: the responsibility that comes with public authority in science. In hindsight, folding this episode into a broader critique of Bell and the Nobel Prize blurred two distinct issues — foundations of physics, and epistemic ethics.

Both matter. They should not be confused.


4. Where I stand now

If there is a single thread connecting my current thinking to these older texts, it is this:

I am less interested than before in winning arguments, and more interested in clarifying where different positions actually part ways — ontologically, methodologically, and institutionally.

That shift is visible elsewhere in my work:

  • in a softer, more discriminating stance toward the Standard Model,
  • in a deliberate break with institutions and labels that locked me into adversarial postures,
  • and in a conscious move toward reconciliation where reconciliation is possible, and clean separation where it is not.

The posts on Bell’s Theorem were written at an earlier stage in that trajectory. I do not disown them. But I no longer want them to stand without context.

This text is that context.


Final notes

1. On method and collaboration

Much of the clarification in this essay did not emerge in isolation, but through extended dialogue — including with an AI interlocutor that acted, at times, less as a generator of arguments than as a moderator of instincts: slowing me down, forcing distinctions, and insisting on separating epistemic claims from emotional charge. That, too, is part of the story — and perhaps an unexpected one. If intellectual maturity means anything, it is not the abandonment of strong positions, but the ability to state them without needing indignation to carry the weight. That is the work I am now trying to do.

It is also why I want to be explicit about how these texts are currently produced: they are not outsourced to AI, but co-generated through dialogue. In that dialogue, I deliberately highlight not only agreements but also remaining disagreements — not on the physics itself, but on its ontological interpretation — with the AI agent I currently use (ChatGPT 5.2). Making those points of convergence and divergence explicit is, I believe, intellectually healthier than pretending they do not exist.

2. On stopping, without pretending to conclude

This post also marks a natural stopping point. Over the past weeks, several long-standing knots in my own thinking — Bell’s Theorem (what this post is about), the meaning of gauge freedom, the limits of Schrödinger’s equation as a model of charge in motion, or even very plain sociological considerations on how sciences moves forward — have either been clarified or cleanly isolated.

What remains most resistant is the problem of matter–antimatter pair creation and annihilation. Here, the theory appears internally consistent, while the experimental evidence, impressive as it is, still leaves a small but non-negligible margin of doubt — largely because of the indirect, assumption-laden nature of what is actually being measured. I do not know the experimental literature well enough to remove that last 5–10% of uncertainty, and I consider it a sign of good mental health not to pretend otherwise.

For now, that is where I leave it. Not as a conclusion, but as a calibration: knowing which questions have been clarified, and which ones deserve years — rather than posts — of further work.

3. Being precise on my use of AI: on cleaning up ideas, not outsourcing thinking

What AI did not do

Let me start with what AI did not do.

It did not:

  • supply new experimental data,
  • resolve open foundational problems,
  • replace reading, calculation, or judgment,
  • or magically dissolve the remaining hard questions in physics.

In particular, it did not remove my residual doubts concerning matter–antimatter pair creation. On that topic, I remain where I have been for some time: convinced that the theory is internally consistent, convinced that the experiments are impressive and largely persuasive, and yet unwilling to erase the remaining 5–10% of doubt that comes from knowing how indirect, assumption-laden, and instrument-mediated those experiments necessarily are. I still do not know the experimental literature well enough to close that last gap—and I consider it a sign of good mental health that I do not pretend otherwise.

What AI did do

What AI did do was something much more modest—and much more useful.

It acted as a moderator of instincts.

In the recent rewrites—most notably in this post (Cleaning Up After Bell)—AI consistently did three things:

  1. It cut through rhetorical surplus.
    Not by softening arguments, but by separating epistemic claims from frustration, indignation, or historical irritation.
  2. It forced distinctions.
    Between mathematical theorems and their physical premises; between experimental closure and ontological interpretation; between criticism of ideas and criticism of institutions.
  3. It preserved the spine while sharpening the blade.
    The core positions did not change. What changed was their articulation: less adversarial, more intelligible, and therefore harder to dismiss.

In that sense, AI did not “correct” my thinking. It helped me re-express it in a way that better matches where I am now—intellectually and personally.

Two primitives or one?

A good illustration is the remaining disagreement between myself and my AI interlocutor on what is ultimately primitive in physics.

I still tend to think in terms of two ontological primitives: charge and fields—distinct, but inseparably linked by a single interaction structure. AI, drawing on a much broader synthesis of formal literature, prefers a single underlying structure with two irreducible manifestations: localized (charge-like) and extended (field-like).

Crucially, this disagreement is not empirical. It is ontological, and currently underdetermined by experiment. No amount of rhetorical force, human or artificial, can settle it. Recognizing that—and leaving it there—is part of intellectual maturity.

Why I am stopping (again)

I have said before that I would stop writing, and I did not always keep that promise. This time, however, the stopping point feels natural.

Most of the conceptual “knots” that bothered me in the contemporary discourse on physics have now been:

  • either genuinely clarified,
  • or cleanly isolated as long-horizon problems requiring years of experimental and theoretical work.

At this point, continuing to write would risk producing more words than signal.

There are other domains that now deserve attention: plain work, family projects, physical activity, and the kind of slow, tangible engagement with the world that no theory—however elegant—can replace.

Closing

If there is a single lesson from this episode, it is this:

AI is most useful not when it gives answers, but when it helps you ask what you are really saying—and whether you still stand by it once the noise is stripped away.

Used that way, it does not diminish thinking.
It disciplines it.

For now, that is enough.

Physics Without Consolations

On Quantum Mechanics, Meaning, and the Limits of Metaphysical Inquiry

This post is a rewritten version of an essay I published on this blog in September 2020 under the title The End of Physics. The original text captured a conviction I still hold: that quantum mechanics is strange but not mysterious, and that much of what is presented as metaphysical depth in modern physics is better understood as interpretive excess. What has changed since then is not the substance of that conviction, but the way I think it should be expressed.

Over the past years, I have revisited several of my physics papers in dialogue with artificial intelligence — not as a replacement for human judgment, but as a tool for clarification, consistency checking, and tone correction. This post is an experiment of the same kind: returning to an older piece of writing with the help of AI, asking not “was I wrong?” but “can this be said more precisely, more calmly, and with fewer rhetorical shortcuts?”

The result is not a repudiation of the 2020 text (and similar ones here on this blog site, or on my ResearchGate page) but a refinement of it.
If there is progress here, it lies not in new claims about physics, but in a clearer separation between what physics tells us about the world and what humans sometimes want it to tell us.

— Jean Louis Van Belle
1 January 2026

After the Mysteries: Physics Without Consolations

For more than a century now, quantum mechanics has been presented as a realm of deep and irreducible mystery. We are told that nature is fundamentally unknowable, that particles do not exist until observed, that causality breaks down at the smallest scales, and that reality itself is somehow suspended in a fog of probabilities.

Yet this way of speaking says more about us than about physics.

Quantum mechanics is undeniably strange. But strange is not the same as mysterious. The equations work extraordinarily well, and — more importantly — we have perfectly adequate physical interpretations for what they describe. Wavefunctions are not metaphysical ghosts. They encode physical states, constraints, and statistical regularities in space and time. Particles such as photons, electrons, and protons are not abstract symbols floating in Hilbert space; they are real physical systems whose behavior can be described using familiar concepts: energy, momentum, charge, field structure, stability.

No additional metaphysics is required.

Over time, however, physics acquired something like a priesthood of interpretation. Mathematical formalisms were promoted from tools to truths. Provisional models hardened into ontologies. Concepts introduced for calculational convenience were treated as if they had to exist — quarks, virtual particles, many worlds — not because experiment demanded it, but because the formalism allowed it.

This is not fraud. It is human behavior.


The Comfort of Indeterminism

There is another, less discussed reason why quantum mechanics became mystified. Indeterminism offered something deeply attractive: a perceived escape hatch from a fully ordered universe.

For some, this meant intellectual freedom. For others, moral freedom. And for some — explicitly or implicitly — theological breathing room.

It is not an accident that indeterminism was welcomed in cultural environments shaped by religious traditions. Many prominent physicists of the twentieth century were embedded — socially, culturally, or personally — in Jewish, Catholic, or Protestant worlds. A universe governed strictly by deterministic laws had long been seen as hostile to divine action, prayer, or moral responsibility. Quantum “uncertainty” appeared to reopen a door that classical physics seemed to have closed.

The institutional embrace of this framing is telling. The Vatican showed early enthusiasm for modern cosmology and quantum theory, just as it did for the Big Bang model — notably developed by Georges Lemaître, a Catholic priest as well as a physicist. The Big Bang fit remarkably well with a creation narrative, and quantum indeterminism could be read as preserving divine freedom in a lawful universe.

None of this proves that physics was distorted intentionally. But it does show that interpretations do not emerge in a vacuum. They are shaped by psychological needs, cultural background, and inherited metaphysical anxieties.


Determinism, Statistics, and Freedom

Rejecting metaphysical indeterminism does not mean endorsing a cold, mechanical universe devoid of choice or responsibility.

Statistical determinism is not fatalism.

Complex systems — from molecules to brains to societies — exhibit emergent behavior that is fully lawful and yet unpredictable in detail. Free will does not require violations of physics; it arises from self-organizing structures capable of evaluation, anticipation, and choice. Moral responsibility is not rescued by randomness. In fact, randomness undermines responsibility far more than lawfulness ever did.

Consciousness, too, does not need mystery to be meaningful. It is one of the most remarkable phenomena we know precisely because it emerges from matter organizing itself into stable, recursive, adaptive patterns. The same principles operate at every scale: atoms in molecules, molecules in cells, cells in organisms, organisms in ecosystems — and, increasingly, artificial systems embedded in human-designed environments.

There is no voice speaking to us from outside the universe. But there is meaning, agency, and responsibility arising from within it.


Progress Without Revelation

It is sometimes said that physics is advancing at an unprecedented pace. In a technical sense, this is true. But conceptually, the situation is more sobering.

Most of the technologies we rely on today — semiconductors, lasers, superconductors, waveguides — were already conceptually understood by the mid-twentieth century and are clearly laid out in The Feynman Lectures on Physics. Later developments refined, scaled, and engineered these ideas, but they did not introduce fundamentally new physical principles.

Large experimental programs have confirmed existing theories with extraordinary precision. That achievement deserves respect. But confirmation is not revelation. Precision is not profundity.

Recognizing this is not pessimism. It is intellectual honesty.


After Physics Ends

If there is an “end of physics,” it is not the end of inquiry, technology, or wonder. It is the end of physics as a source of metaphysical consolation. The end of physics as theology by other means.

What remains is enough: a coherent picture of the material world, an understanding of how complexity and consciousness arise, and the responsibility that comes with knowing there is no external guarantor of meaning.

We are on our own — but not lost.

And that, perhaps, is the most mature scientific insight of all.

One Equation, Too Many Jobs: Rethinking Schrödinger’s Equation and Wavefunction

I have just republished one of my long-standing papers on de Broglie’s matter-wave concept as a new, standalone publication, with its own DOI:

👉 De Broglie’s matter-wave concept and issues
https://www.researchgate.net/publication/399225854_De_Broglie’s_matter-wave_concept_and_issues
DOI: 10.13140/RG.2.2.30104.25605

The reason for republishing is not cosmetic. A new Annex was added on 31 December 2025 that fundamentally clarified — for me, at least — what Schrödinger’s equation is really doing, and just as importantly, what it is not doing.

This clarification came out of a long and at times uncomfortable dialogue with the most recent version of OpenAI’s GPT model (ChatGPT 5.2). Uncomfortable, because it initially destabilized a view I had held for years. Productive, because it forced a deeper structural distinction that I now believe is unavoidable. Let me explain.


The uncomfortable admission: I was wrong about the 12\tfrac{1}{2} factor​

For a long time, I was convinced that the factor 12\tfrac{1}{2} factor in Schrödinger’s equation — especially in the hydrogen atom problem — must reflect some deeper pairing mechanism. At times, I even wondered whether the equation was implicitly modeling an electron pair (opposite spin), rather than a single electron.

That intuition was not random. It came from a broader realist programme in which I treat the electron as a structured object, with internal dynamics (zitterbewegung-like orbital motion), not as a point particle. If mass, energy, and phase all have internal structure, why should a simple quadratic kinetic term with a mysterious 12\tfrac{1}{2} be fundamental?

The hard truth is this: that intuition was misplaced — but it was pointing in the right direction.

The mistake was not questioning the factor 12\tfrac{1}{2}. The mistake was assuming Schrödinger’s equation was trying to describe everything at once.


The key insight: Schrödinger describes the envelope, not the engine

The decisive realization was structural:

Schrödinger’s wavefunction does not describe the electron’s internal dynamics.
It describes the translational envelope of phase coherence.

Once you see that, several things fall into place immediately:

  • The hydrogen “orbitals” are not literal orbits, and not internal electron motion.
  • They are standing-wave solutions of an envelope phase, constrained by a Coulomb potential.
  • The factor 12\tfrac{1}{2}​ is not mysterious at all at this level: it is the natural coefficient that appears in effective, averaged, quadratic envelope dynamics.

In other words:
The 12\tfrac{1}{2} factor belongs to the envelope layer, not to the internal structure of the electron.

My earlier “electron pair” idea tried to explain a structural feature by inventing new ontology. The correct move was simpler and more radical: separate the layers.


One symbol, too many jobs

Modern quantum mechanics makes a profound — and in my view costly — simplification:

It uses one symbol, ψ, to represent:

  • internal phase,
  • translational dynamics,
  • probability amplitudes,
  • and experimental observables.

That compression works operationally, but it hides structure.

What the new Annex makes explicit is that Nature almost certainly does not work that way. At minimum, we should distinguish:

  1. Internal phase
    Real, physical, associated with internal orbital motion and energy bookkeeping.
  2. Envelope phase
    Slow modulation across space, responsible for interference, diffraction, and spectra.
  3. Observables
    What experiments actually measure, which are sensitive mainly to envelope-level phase differences.

Once this distinction is made, long-standing confusions dissolve rather than multiply.


Why this does not contradict experiments

This is crucial.

Nothing in this reinterpretation invalidates:

  • electron diffraction,
  • hydrogen spectra,
  • interference experiments,
  • or the empirical success of standard quantum mechanics.

On the contrary: it explains why Schrödinger’s equation works so well — within its proper domain.

The equation is not wrong.
It is just over-interpreted.


A personal note on changing one’s mind

I’ll be honest: this line of reasoning initially felt destabilizing. It challenged a position I had defended for years. But that discomfort turned out to be a feature, not a bug.

Good theory-building does not preserve intuitions at all costs. It preserves structure, coherence, and explanatory power.

What emerged is a cleaner picture:

  • internal realism without metaphysics,
  • Schrödinger demoted from “ultimate truth” to “effective envelope theory”,
  • and a much clearer map of where different mathematical tools belong.

That, to me, is progress.


Where this opens doors

Once we accept that one wavefunction cannot represent all layers of Nature, new possibilities open up:

  • clearer interpretations of spin and the Dirac equation,
  • better realist models of lattice propagation,
  • a more honest treatment of “quantum mysteries” as category mistakes,
  • and perhaps new mathematical frameworks that respect internal structure from the start.

Those are not promises — just directions.

For now, I am satisfied that one long-standing conceptual knot has been untied.

And sometimes, that’s enough for a good year’s work. 🙂

Post Scriptum: On AI, Intellectual Sparring, and the Corridor

A final remark, somewhat orthogonal to physics.

The revision that led to this blog post and the accompanying paper did not emerge from a sudden insight, nor from a decisive experimental argument. It emerged from a long, occasionally uncomfortable dialogue with an AI system, in which neither side “won,” but both were forced to refine their assumptions.

At the start of that dialogue, the AI responded in a largely orthodox way, reproducing standard explanations for the factor 12\tfrac{1}{2}​ in Schrödinger’s equation. I, in turn, defended a long-held intuition that this factor must point to internal structure or pairing. What followed was not persuasion, but sparring: resistance on both sides, followed by a gradual clarification of conceptual layers. The breakthrough came when it became clear that a single mathematical object — the wavefunction — was being asked to do too many jobs at once.

From that moment on, the conversation shifted from “who is right?” to “which layer are we talking about?” The result was not a victory for orthodoxy or for realism, but a structural separation: internal phase versus translational envelope, engine versus modulation. That separation resolved a tension that had existed for years in my own thinking.

I have explored this mode of human–AI interaction more systematically in a separate booklet on ResearchGate, where I describe such exchanges as occurring within a corridor: a space in which disagreement does not collapse into dominance or deference, but instead forces both sides toward finer distinctions and more mature reasoning.

This episode convinced me that the real intellectual value of AI does not lie in answers, but in sustained resistance without ego — and in the willingness of the human interlocutor to tolerate temporary destabilization without retreating into dogma. When that corridor holds, something genuinely new can emerge.

In that sense, this post is not only about Schrödinger’s equation. It is also about how thinking itself may evolve when humans and machines are allowed to reason together, rather than merely agree.

Readers interested in this kind of human–AI interaction beyond the present physics discussion may want to look at that separate booklet I published on ResearchGate (≈100 pages), in which I try to categorize different modes of AI–human intellectual interaction — from superficial compliance and authority projection to genuine sparring. In that text, exchanges like the one briefly alluded to above are described as a Type-D collapse: a situation in which both human and AI are forced to abandon premature explanatory closure, without either side “winning,” and where progress comes from structural re-layering rather than persuasion.

The booklet is intentionally exploratory and occasionally playful in tone, but it grew out of exactly this kind of experience: moments where resistance, rather than agreement, turns out to be the most productive form of collaboration.

Perhaps we will stop here – time will tell :-)

I have just uploaded a new working paper to ResearchGate: Ontology, Physics, and Math – Einstein’s Unfinished Revolution. I am not announcing it with any sense of urgency, nor with the expectation that it will “change” physics. If it contributes anything at all, it may simply offer a bit of clarity about what we can reasonably claim to see in physics — and what we merely calculate, fit, or postulate. That distinction has preoccupied me for years.

A space to think

One unexpected consequence of taking AI seriously over the past one or two years is that it restored something I had quietly lost: a space to think.

  • Not a space to produce.
  • Not a space to publish.
  • Not a space to compete.

Just a space to think — slowly, carefully, without having to defend a position before it has fully formed. That kind of space has become rare. Academia is under pressure, industry is under pressure, and even independent thinkers often feel compelled to rush toward closure. The conversations I’ve had with AI — what I’ve come to call a corridor — were different. They were not about winning arguments, but about keeping the corridor open only where conceptual clarity survived.

In a strange way, this brought me back to something much older than AI. When I was young, I wanted to study philosophy. My father refused. I had failed my mathematics exam for engineering studies, and in his view philosophy without mathematics was a dead end. In retrospect, I can see that he was probably right — and also that he struggled with me as much as I struggled with him. He should perhaps have pushed me into mathematics earlier; I should perhaps have worked harder. But life does not run backward, and neither does understanding. What AI unexpectedly gave me, decades later, was the chance to reunite those two threads: conceptual questioning disciplined by mathematical restraint. Not philosophy as free-floating speculation, and not mathematics as pure formalism — but something closer to what physics once called natural philosophy.

Why I was always uncomfortable

For a long time, I could not quite place my discomfort. I was uneasy with mainstream Standard Model theorists — not because their work lacks brilliance or empirical success (it clearly does not), but because formal success increasingly seemed to substitute for ontological clarity. At the same time, I felt equally uneasy among outsiders and “fringe” thinkers, who were often too eager to replace one elaborate ontology with another, convinced that the establishment had simply missed the obvious.

I now think I understand why I could not belong comfortably to either camp. Both, in different ways, tend to underestimate what went into building the Standard Model in the first place.

  • The Standard Model is not just a theory. It is the result of enormous societal investment (yes, taxes matter), decades of engineering ingenuity, and entire academic ecosystems built around measurement, refinement, and internal consistency. One does not wave that away lightly. Criticizing it without acknowledging that effort is not radical — it is careless.
  • At the same time, acknowledging that effort does not oblige one to treat the resulting ontology as final. Formal closure is not the same thing as physical understanding.

That tension — respect without reverence — is where I found myself stuck.

Seeing versus calculating

The paper I just uploaded does not attempt to overthrow the Standard Model, nor to replace ΛCDM, nor to propose a new unification. It does something much more modest: it tries to separate what we can physically interpret from what we can formally manipulate.

That distinction was central to the worries of people like Albert Einstein, long before it became unfashionable to worry about such things. Einstein’s famous remark to Max Born“God does not play dice” — was not a rejection of probability as a calculational tool. It was an expression of discomfort with mistaking a formalism for a description of reality. Something similar motivated Louis de Broglie, and later thinkers who never quite accepted that interpretation should be outsourced entirely to mathematics.

What my paper argues — cautiously, and without claiming finality — is that much of modern physics suffers from a kind of ontological drift: symmetries that began life as mathematical operations sometimes came to be treated as physical mandates.

When those symmetries fail, new quantum numbers, charges, or conservation laws are introduced to restore formal order. This works extraordinarily well — but it also risks confusing bookkeeping with explanation.

Matter, antimatter, and restraint

The most difficult part of the paper concerns matter–antimatter creation and annihilation. For a long time, I resisted interpretations that treated charge as something that could simply appear or disappear. That resistance did not lead me to invent hidden reservoirs or speculative intermediates — on the contrary, I explicitly rejected such moves as ontological inflation. Instead, I left the tension open.

Only later did I realize that insisting on charge as a substance may itself have been an unjustified metaphor. Letting go of that metaphor did not solve everything — but it did restore coherence without adding entities. That pattern — refusing both cheap dismissal and cheap solutions — now feels like the right one.

Ambition, patience, and time

We live in a period of extraordinary measurement and, paradoxically, diminished understanding. Data accumulates. Precision improves. Parameters are refined. But the underlying picture often becomes more fragmented rather than more unified.

New machines may or may not be built. China may or may not build the next CERN. That is largely beyond the control of individual thinkers. What is within reach is the slower task of making sense of what we already know. That task does not reward ambition. It rewards patience.

This is also where I part ways — gently, but firmly — with some bright younger thinkers and some older, semi-wise ones. Not because they are wrong in detail, but because they sometimes underestimate the weight of history, infrastructure, and collective effort behind the theories they critique or attempt to replace. Time will tell whether their alternatives mature. Time always tells :-). […] PS: I add a ‘smiley’ here because, perhaps, that is the most powerful phrase of all in this post.

A pause, not a conclusion

This paper may mark the end of my own physics quest — or at least a pause. Not because everything is resolved, but because I finally understand why I could neither fully accept nor fully reject what I was given. I don’t feel compelled anymore to choose sides. I can respect the Standard Model without canonizing it, and I can question it without trying to dethrone it. I can accept that some questions may remain open, not because we lack data, but because clarity sometimes requires restraint.

For now, that feels like enough. Time to get back on the bike. 🙂

PS: Looking back at earlier philosophical notes I wrote years ago — for instance on the relation between form, substance, and charge — I’m struck less by how “wrong” they were than by how unfinished they remained. The questions were already there; what was missing was discipline. Not more speculation, but sharper restraint.

We Could Have Stopped There Too

(But the Question About Annihilation Would Not Stay Quiet)

In a previous post, I wrote that we could stop here — after revisiting the photon wavefunction and trying to say, as carefully as possible, what such a wavefunction might represent in physical reality rather than merely in calculation. That paper already felt like a natural resting point: the mathematics was consistent, the interpretation restrained, and the temptation to add speculative layers had been resisted.

But, as often happens, the very act of stopping made the next question louder.

If one is willing to take wavefunctions seriously — not as mystical probability clouds but as structured representations of physical processes — then one cannot avoid revisiting an older and more uncomfortable puzzle: matter–antimatter pair creation and annihilation. In particular, the question that has bothered me for years refused to go away:

What, exactly, happens to electric charge in electron–positron annihilation?

In January 2025, I wrote a paper on this topic together with ChatGPT-4.0. That version deliberately stopped short of resolution. It explored wavefunctional representations, respected global conservation laws, and openly admitted that familiar intuitions about charge seemed to fail locally. I resisted easy exits: latent charge states, hidden reservoirs, or metaphysical bookkeeping devices introduced only to preserve comfort.

At the time, that felt honest enough.

What changed since then is not the question, but the discipline with which I was forced to re-examine my own assumptions.

Over the past months, continued work with a more advanced AI system (ChatGPT-5.2), across many iterations and with partial memory of prior discussions, introduced a form of pressure that was unfamiliar but productive. The AI did not argue for a competing ontology. Instead, it kept doing something more unsettling: it repeatedly asked why certain assumptions were still being carried along at all.

In hindsight, I can see that I was still clinging — subconsciously — to the idea that charge must be something that persists, even if I no longer knew where to put it. That assumption had survived earlier criticism not because it was well-justified, but because it was deeply ingrained.

What finally shifted the balance was a stricter application of Occam’s razor — applied not to equations, but to ontological commitments. If charge is inseparable from a specific physical organization (of motion, phase, and localization), then insisting that it must survive the dissolution of that organization is not conservative reasoning. It is surplus.

This led, reluctantly but unavoidably, to a provisional reformulation: perhaps charge is not a substance that must “go somewhere,” but a mode of organization that ceases to exist when the organization itself dissolves. This idea is not offered as a new metaphysical doctrine. On the contrary, it emerged as a refusal to introduce additional entities whose only role would be to save intuition.

The revised paper therefore appears in two parts. The January version is preserved intact, as a record of where the reasoning stood at that time. The new December revision does not correct it so much as re-read it under harsher criteria of conceptual economy. Several distinctions — including the boson–fermion divide — remain descriptively useful, but are relieved of explanatory burdens they were never meant to carry.

As before, no final answers are claimed. The ontological and philosophical implications are intentionally left for the reader — real or imaginary — to judge. The role of AI in this process was not to supply insight, but to apply relentless pressure against conceptual inertia. Any logical errors or unwarranted commitments that remain are mine alone, even if much of the textual consistency was produced by artificial means.

We could, perhaps, stop here as well.

But I have learned to be suspicious of that feeling. When a question keeps knocking, it is usually because something unnecessary is still being held onto — and is asking to be let go.

We Could Stop Here.

(But the Next Question Is Already Knocking.)

There is a moment in any long intellectual journey where you could stop.

Not because everything is finished, but because enough has settled to make stopping respectable. The equations close. The concepts line up. Nothing is obviously broken anymore.

This paper — The Photon Wavefunction Revisited — marks one of those moments for me.

👉 The paper is available here on ResearchGate:
https://www.researchgate.net/publication/399111974_The_Photon_Wavefunction_Revisited

It revisits an old and stubborn question — what do we really mean by the photon wavefunction? — using only very old tools: Maxwell’s equations, the Planck–Einstein relation, dimensional analysis, and known scattering results. No new particles. No speculative fields. No hidden dimensions. No “next revolution”.

Just careful rereading.

Why revisit this at all?

Because physics has a habit of answering questions so efficiently that we stop asking what the answers mean. The photon became a “quantum of the electromagnetic field”, calculations worked, experiments agreed — and interpretation quietly retreated.

But interpretation has a way of sneaking back in through the side door.

In this paper, I try to be very explicit about what is being claimed — and what is not:

  • A photon is treated as a light-like, phase-closed object, not as a little billiard ball and not as a probabilistic smear.
  • Its wavefunction is not a mystery object “without meaning”, but a compact encoding of phase structure.
  • Electric and magnetic fields are not competing realities, but orthogonal phase components of a single conserved structure.
  • Energy and momentum conservation follow cleanly from Maxwell’s equations — even when charge is stripped away.

Nothing here overturns quantum electrodynamics. But some things are, perhaps, put back in their original place.

A word about standing waves (and why they appear)

One appendix uses a standing-wave construction to make something visible that is otherwise hidden: how electric and magnetic field energy exchange internally while total energy remains conserved.

This does not mean photons are standing waves. They propagate in one direction. Momentum has a direction. Energy does not.

The standing wave is simply a diagnostic tool — a way of freezing momentum flow so the bookkeeping of energy becomes transparent. If that sounds almost embarrassingly classical… well, that may be the point.

Why this felt worth publishing

This paper took shape slowly, through many iterations, many dead ends, and many “wait — is that actually true?” moments. Some of it was developed with explicit AI assistance, used not as an oracle but as a very patient consistency checker. That role is openly acknowledged.

What mattered most to me was not novelty, but coherence.

When the dust settled, something quietly reassuring happened: the picture that emerged was simpler than what I started with, not more complicated.

And that’s usually a good sign.

Could we stop here?

Yes. Absolutely.

The paper stands on its own. The equations close. Nothing essential is missing.

But physics has never progressed by stopping at “good enough”. The next question is already there:

  • How exactly does this phase picture illuminate electron–photon interaction?
  • What does it really say about the fine-structure constant?
  • Where does this leave matter–antimatter symmetry?

Those are not answered here. They don’t need to be — yet.

For now, this is a place to pause, look around, and make sure we know where we are.

And then, as always, the next question prompts the next question.

That’s not a problem.
That’s the fun part.

— Jean Louis Van Belle

Post Scriptum: The Last Question That Won’t Let Me Sleep (On matter, antimatter, and why one mystery remains)

There is a strange pattern I’ve noticed over the years.

You work your way through a dense thicket of questions. One by one, they loosen. Concepts that once felt contradictory begin to align. The mathematics stops fighting the intuition. The ontology — cautiously, provisionally — starts to hold.

And then, when almost everything is in place, one question refuses to dissolve.

Tonight, for me, that question is matter–antimatter creation and annihilation.

Most things now feel… settled

After revisiting photons, wavefunctions, phase closure, and electromagnetic energy bookkeeping, I feel unusually calm about many things that once bothered me deeply.

  • Photons as light-like, phase-closed objects? That works.
  • Electric and magnetic fields as orthogonal phase components? That works.
  • Energy conservation without charge? Maxwell already knew how to do that.
  • Electron–photon interaction as phase reconfiguration rather than “mystical coupling”? That works too.

None of this feels revolutionary anymore. It feels readable.

And yet.

Matter–antimatter still feels different

In low-energy environments, I’m increasingly comfortable with a very unromantic picture.

Pair creation does not happen “out of nothing.” It happens near nuclei, in strong fields, in structured environments. Something must anchor phase. Something must absorb recoil. Something must allow a stable oscillatory configuration to form.

I’ve sometimes called this a Platzwechsel — a change of place, or role — rather than a miraculous transformation of field into charge. The photon doesn’t “become matter”; a charge configuration re-closes in the presence of structure.

That feels honest. And it fits what experiments actually show.

But then there is the “but” question… This is how I phrase now.

Annihilation is unsettlingly easy

Electron–positron annihilation, on the other hand, requires no such help.

Two charged, massive objects meet, and they disappear into light. Cleanly. Elegantly. No nucleus. No lattice. No scaffold.

That asymmetry matters.

Matter → light is easy.
Light → matter is hard.

Quantum field theory encodes this perfectly well, but encoding is not explaining. And pretending the asymmetry isn’t there has never helped.

What happens to charge?

Here is the thought that keeps me awake — and oddly calm at the same time.

If charge is not a substance, but a phase-closed electromagnetic motion, then annihilation is not mysterious at all. The phase closure simply dissolves. What remains is free phase propagation.

Charge doesn’t “go anywhere”.
It stops being a thing because the structure that constituted it no longer exists.

That idea is unsettling only if one insists that charge must persist locally as a substance. I’ve never found good reasons to believe that.

And pure vacuum pair creation?

High-energy photon–photon pair creation is possible, in principle. But it is rare, fragile, and structurally demanding. It requires extreme energies and densities, and often still some form of external assistance.

That, too, feels telling.

Two freely propagating phase objects have no natural way to decide where a charge configuration should live. Without structure, closure is unstable. Nature seems reluctant — not forbidden, but reluctant.

So where does that leave us?

It leaves me in an oddly peaceful place.

Most of the framework now feels coherent. The remaining mystery is not a loose end to be tied up quickly, but a boundary — a place where explanation must slow down instead of speeding up.

That feels like the right place to stop for tonight.

Not because the mystery is solved, but because it is now cleanly stated.

And that, I’ve learned, is often the real precondition for sleep.

— Jean Louis Van Belle

When Decay Statistics Become Ontology

Or: why the Standard Model feels so solid — and yet so strangely unsatisfying

I recently put a new paper online: A Taxonomy of Instability. It is, in some sense, a “weird” piece. Not because it proposes new particles, forces, or mechanisms — it does none of that — but because it deliberately steps sideways from the usual question:

What are particles made of?

and asks instead:

How do unstable physical configurations actually fail?

This shift sounds modest. In practice, it leads straight into a conceptual fault line that most of us sense, but rarely articulate.


What is actually being classified in particle physics?

The Standard Model is extraordinarily successful. That is not in dispute. It predicts decay rates, cross sections, and branching fractions with astonishing precision. It has survived decades of experimental scrutiny.

But it is worth noticing what it is most directly successful at describing:

  • lifetimes,
  • branching ratios,
  • observable decay patterns.

In other words: statistics of instability.

Yet when we talk about the Standard Model, we almost immediately slide from that statistical success into an ontological picture: particles as entities with intrinsic properties, decaying “randomly” according to fundamental laws.

That slide is so familiar that it usually goes unnoticed.


The quiet assumption we almost never examine

Consider how decay is presented in standard references (PDG tables are the cleanest example). For a given unstable particle, we are shown:

  • a list of decay “channels”,
  • each with a fixed branching fraction,
  • averaged over production mechanisms, environments, and detectors.

Everything contextual has been stripped away.

What remains is treated as intrinsic.

And here is where a subtle but radical assumption enters:

The same unstable particle is taken to be capable of realizing multiple, structurally distinct decay reactions, with no further individuation required.

This is not an experimental result.
It is an interpretive stance.

As long as one stays in calculational mode, this feels unproblematic. The formalism works. The predictions are right.

The discomfort only arises when one asks a very basic question:

If all environment variables are abstracted away, what exactly is it that is decaying?


Statistical determinism sharpens the problem

Decay statistics are not noisy or unstable. They are:

  • reproducible,
  • environment-independent (within stated limits),
  • stable across experiments.

That makes them look law-like.

But law-like behavior demands clarity about what level of description the law applies to.

There are two logically distinct possibilities:

  1. Intrinsic multivalence
    A single physical entity genuinely has multiple, mutually exclusive decay behaviors, realized stochastically, with no deeper individuation.
  2. Hidden population structure
    What we call “a particle” is actually an equivalence class of near-identical configurations, each with a preferred instability route, unresolved by our current classification.

The Standard Model chooses option (1) — implicitly, pragmatically, and very effectively.

But nothing in the data forces that choice.


Why this can feel like being “duped”

Many people only experience discomfort after they start thinking carefully about what the Standard Model is claiming to describe.

The sense of being “duped” does not come from experimental failure — it comes from realizing that a philosophical commitment was made silently, without being labeled as such.

Probability, in this framework, is not treated as epistemic (what we don’t know), but as ontologically primitive (what is). Identity is divorced from behavior. The ensemble description quietly replaces individual determinism.

This is a perfectly legitimate move — but it is a move.

And it has a cost.


What my taxonomy does — and does not — claim

A Taxonomy of Instability does not propose new physics. It does not challenge the predictive success of the Standard Model. It does not deny quantum mechanics.

What it does is much quieter:

  • it treats decay landscapes, not particles, as the primary objects of classification;
  • it groups unstable configurations by how they fail, not by assumed internal structure;
  • it keeps the description strictly operational: lifetimes, observable final states, branching structure.

In doing so, it exposes something we usually gloss over:

Treating statistically distinct instability morphologies as attributes of a single identity is already an ontological decision.

Once that decision is made explicit, it becomes optional rather than compulsory.


Why this feels “weird” — and why that’s a good sign

The paper feels strange because it does not do what most theoretical work does:

  • it does not explain,
  • it does not unify,
  • it does not speculate about deeper mechanisms.

Instead, it asks whether our classification layer has quietly hardened into ontology.

That kind of question always feels uncomfortable, because it sits between theory and philosophy, and because it removes a tacit compromise rather than proposing a new belief.

But it is also the kind of question that matters precisely when a theory works extremely well.


A broader resonance (human and artificial)

There is an additional reason this question feels timely.

Modern AI systems are, at their core, pattern classifiers and compressors. They turn data into “things” by grouping outcomes under labels. Ontologies emerge automatically unless we are careful.

Seen from that angle, particle physics is not an outlier — it is an early, highly successful example of how statistical regularities become reified as entities.

The taxonomy I propose is not only about particles. It is about how thinking systems — human or artificial — turn data into objects.


A calm conclusion

The Standard Model is an extraordinarily successful theory of decay statistics. Its difficulties are not primarily empirical, but philosophical.

Those difficulties arise only when we forget that:

  • classification is not explanation,
  • identity is not forced by statistics,
  • and ontology is not delivered for free by predictive success.

My hope is not to replace any existing framework, but to invite both human readers and artificial “thinking machines” to pause and ask again:

What is being measured — and what, exactly, are we saying exists?

Sometimes, the most productive form of progress is not adding a new layer, but noticing where an old one quietly became invisible.

Stability First: A Personal Programme for Re-reading Particle Physics

Over the past years, I have written a number of papers on physics—mostly exploratory, sometimes speculative, always driven by the same underlying discomfort.

Not with the results of modern physics. Those are extraordinary.
But with the ordering of its explanations.

We are very good at calculating what happens.
We are less clear about why some things persist and others do not.

That question—why stability appears where it does—has quietly guided much of my thinking. It is also the thread that ties together a new manuscript I have just published on ResearchGate:

“Manuscript v0.2 – A stability-first reinterpretation of particle physics”
👉 https://www.researchgate.net/publication/398839393_Manuscript_v02

This post is not a summary of the manuscript. It is an explanation of why I wrote it, and what kind of work it is meant to enable.


Not a new theory — a different starting point

Let me be clear from the outset.

This manuscript does not propose a new theory.
It does not challenge the empirical success of the Standard Model.
It does not attempt to replace quantum field theory or nuclear phenomenology.

What it does is much more modest—and, I hope, more durable.

It asks whether we have been starting our explanations at the wrong end.

Instead of beginning with abstract constituents and symmetries, the manuscript begins with something far more pedestrian, yet physically decisive:

Persistence in time.

Some entities last.
Some decay.
Some exist only fleetingly as resonances.
Some are stable only in the presence of others.

Those differences are not cosmetic. They shape the physical world we actually inhabit.


From electrons to nuclei: stability as a guide

The manuscript proceeds slowly and deliberately, revisiting familiar ground:

  • the electron, as an intrinsically stable mode;
  • the proton, as a geometrically stable but structurally richer object;
  • the neutron, as a metastable configuration whose stability exists only in relation;
  • the deuteron, as the simplest genuinely collective equilibrium;
  • and nuclear matter, where stability becomes distributed across many coupled degrees of freedom.

At no point is new empirical content introduced.
What changes is the interpretive emphasis.

Stability is treated not as an afterthought, but as a physical clue.


Interaction without mysticism

The same approach is applied to interaction.

Scattering and annihilation are reinterpreted not as abstract probabilistic events, but as temporary departures from equilibrium and mode conversion between matter-like and light-like regimes.

Nothing in the standard calculations is altered.
What is altered is the physical picture.

Wavefunctions remain indispensable—but they are treated as representations of physical configurations, not as substitutes for them.

Probability emerges naturally from limited access to phase, geometry, and configuration, rather than from assumed ontological randomness.


Why classification matters

The manuscript ultimately turns to the Particle Data Group catalogue.

The PDG tables are one of the great achievements of modern physics. But they are optimized for calculation, not for intuition about persistence.

The manuscript proposes a complementary, stability-first index of the same data:

  • intrinsically stable modes,
  • metastable particle modes,
  • prompt decayers,
  • resonances,
  • and context-dependent stability (such as neutrons in nuclei).

Nothing is removed.
Nothing is denied.

The proposal is simply to read the catalogue as a map of stability regimes, rather than as a flat ontology of “fundamental particles”.


A programme statement, not a conclusion

This manuscript is intentionally incomplete.

It does not contain the “real work” of re-classifying the entire PDG catalogue. That work lies ahead and will take time, iteration, and—no doubt—many corrections.

What the manuscript provides is something else:

a programme statement.

A clear declaration of what kind of questions I think are still worth asking in particle physics, and why stability—rather than constituent bookkeeping—may be the right place to ask them from.


Why I am sharing this now

I am publishing this manuscript not as a final product, but as a marker.

A marker of a line of thought I intend to pursue seriously.
A marker of a way of reading familiar physics that I believe remains underexplored.
And an invitation to discussion—especially critical discussion—on whether this stability-first perspective is useful, coherent, or ultimately untenable.

Physics progresses by calculation.
It matures by interpretation.

This manuscript belongs to the second category.

If that resonates with you, you may find the full text of interest.


Jean-Louis Van Belle
readingfeynman.org

Something Rotten in the State of QED? A Careful Look at Critique, Sociology, and the Limits of Modern Physics

Every few years, a paper comes along that stirs discomfort — not because it is wrong, but because it touches a nerve.
Oliver Consa’s Something is rotten in the state of QED is one of those papers.

It is not a technical QED calculation.
It is a polemic: a long critique of renormalization, historical shortcuts, convenient coincidences, and suspiciously good matches between theory and experiment. Consa argues that QED’s foundations were improvised, normalized, mythologized, and finally institutionalized into a polished narrative that glosses over its original cracks.

This is an attractive story.
Too attractive, perhaps.
So instead of reacting emotionally — pro or contra — I decided to dissect the argument with a bit of help.

At my request, an AI language model (“Iggy”) assisted in the analysis. Not to praise me. Not to flatter Consa. Not to perform tricks.
Simply to act as a scalpel: cold, precise, and unafraid to separate structure from rhetoric.

This post is the result.


1. What Consa gets right (and why it matters)

Let’s begin with the genuinely valuable parts of his argument.

a) Renormalization unease is legitimate

Dirac, Feynman, Dyson, and others really did express deep dissatisfaction with renormalization. “Hocus-pocus” was not a joke; it was a confession.

Early QED involved:

  • cutoff procedures pulled out of thin air,
  • infinities subtracted by fiat,
  • and the philosophical hope that “the math will work itself out later.”

It did work out later — to some extent — but the conceptual discomfort remains justified. I share that discomfort. There is something inelegant about infinities everywhere.

b) Scientific sociology is real

The post-war era centralized experimental and institutional power in a way physics had never seen. Prestige, funding, and access influenced what got published and what was ignored. Not a conspiracy — just sociology.

Consa is right to point out that real science is messier than textbook linearity.

c) The g–2 tension is real

The ongoing discrepancy between experiment and the Standard Model is not fringe. It is one of the defining questions in particle physics today.

On these points, Consa is a useful corrective:
he reminds us to stay honest about historical compromises and conceptual gaps.


2. Where Consa overreaches

But critique is one thing; accusation is another.

Consa repeatedly moves from:

“QED evolved through trial and error”
to
“QED is essentially fraud.”

This jump is unjustified.

a) Messiness ≠ manipulation

Early QED calculations were ugly. They were corrected decades later. Experiments did shift. Error bars did move.

That is simply how science evolves.

The fact that a 1947 calculation doesn’t match a 1980 value is not evidence of deceit — it is evidence of refinement. Consa collapses that distinction.

b) Ignoring the full evidence landscape

He focuses almost exclusively on:

  • the Lamb shift,
  • the electron g–2,
  • the muon g–2.

Important numbers, yes — but QED’s experimental foundation is vastly broader:

  • scattering cross-sections,
  • vacuum polarization,
  • atomic spectra,
  • collider data,
  • running of α, etc.

You cannot judge an entire theory on two or three benchmarks.

c) Underestimating theoretical structure

QED is not “fudge + diagrams.”
It is constrained by:

  • Lorentz invariance,
  • gauge symmetry,
  • locality,
  • renormalizability.

Even if we dislike the mathematical machinery, the structure is not arbitrary.

So: Consa reveals real cracks, but then paints the entire edifice as rotten.
That is unjustified.


3. A personal aside: the Zitter Institute and the danger of counter-churches

For a time, I was nominally associated with the Zitter Institute — a loosely organized group exploring alternatives to mainstream quantum theory, including zitterbewegung-based particle models.

I now would like to distance myself.

Not because alternative models are unworthy — quite the opposite. But because I instinctively resist:

  • strong internal identity,
  • suspicion of outsiders,
  • rhetorical overreach,
  • selective reading of evidence,
  • and occasional dogmatism about their own preferred models.

If we criticize mainstream physics for ad hoc factors, we must be brutal about our own.

Alternative science is not automatically cleaner science.


4. Two emails from 2020: why good scientists can’t always engage

This brings me to two telling exchanges from 2020 with outstanding experimentalists: Prof. Randolf Pohl (muonic hydrogen) and Prof. Ashot Gasparian (PRad).

Both deserve enormous respect, and I won’t reveal the email exchanges because of respect, GDPR rules or whatever).
Both email exchanges revealed the true bottleneck in modern physics to me — it is not intelligence, not malice, but sociology and bandwidth.

a) Randolf Pohl: polite skepticism, institutional gravity

Pohl was kind but firm:

  • He saw the geometric relations I proposed as numerology.
  • He questioned applicability to other particles.
  • He emphasized the conservatism of CODATA logic.

Perfectly valid.
Perfectly respectable.
But also… perfectly bound by institutional norms.

His answer was thoughtful — and constrained.
(Source: ChatGPT analysis of emails with Prof Dr Pohl)

b) Ashot Gasparian: warm support, but no bandwidth

Gasparian responded warmly:

  • “Certainly your approach and the numbers are interesting.”
  • But: “We are very busy with the next experiment.”

Also perfectly valid.
And revealing:
even curious, open-minded scientists cannot afford to explore conceptual alternatives.

Their world runs on deadlines, graduate students, collaborations, grants.

(Source: ChatGPT analysis of emails with Prof Dr Pohl)

The lesson

Neither professor dismissed the ideas because they were nonsensical.
They simply had no institutional space to pursue them.

That is the quiet truth:
the bottleneck is not competence, but structure.


5. Why I now use AI as an epistemic partner

This brings me to the role of AI.

Some colleagues (including members of the Zitter Institute) look down on using AI in foundational research. They see it as cheating, or unserious, or threatening to their identity as “outsiders.”

But here is the irony:

AI is exactly the tool that can think speculatively without career risk.

An AI:

  • has no grant committee,
  • no publication pressure,
  • no academic identity to defend,
  • no fear of being wrong,
  • no need to “fit in.”

That makes it ideal for exploratory ontology-building.

Occasionally, as in the recent paper I co-wrote with Iggy — The Wonderful Theory of Light and Matter — it becomes the ideal partner:

  • human intuition + machine coherence,
  • real-space modeling without metaphysical inflation,
  • EM + relativity as a unified playground,
  • photons, electrons, protons, neutrons as geometric EM systems.

This is not a replacement for science.
It is a tool for clearing conceptual ground,
where overworked, over-constrained academic teams cannot go.


6. So… is something rotten in QED?

Yes — but not what you think.

What’s rotten is the mismatch

between:

  • the myth of QED as a perfectly clean, purely elegant theory,
    and
  • the reality of improvised renormalization, historical accidents, social inertia, and conceptual discomfort.

What’s rotten is not the theory itself,
but the story we tell about it.

What’s not rotten:

  • the intelligence of the researchers,
  • the honesty of experimentalists,
  • the hard-won precision of modern measurements.

QED is extraordinary.
But it is not infallible, nor philosophically complete, nor conceptually finished.

And that is fine.

The problem is not messiness.
The problem is pretending that messiness is perfection.


7. What I propose instead

My own program — pursued slowly over many years — is simple:

  • Bring physics back to Maxwell + relativity as the foundation.
  • Build real-space geometrical models of all fundamental particles.
  • Reject unnecessary “forces” invented to patch conceptual holes.
  • Hold both mainstream and alternative models to the same standard:
    no ad hoc constants, no magic, no metaphysics.

And — unusually —
use AI as a cognitive tool, not as an oracle.

Let the machine check coherence.
Let the human set ontology.

If something emerges from the dialogue — good.
If not — also good.

But at least we will be thinking honestly again.


Conclusion

Something is rotten in the state of QED, yes —
but the rot is not fraud or conspiracy.

It is the quiet decay of intellectual honesty behind polished narratives.

The cure is not shouting louder, or forming counter-churches, or romanticizing outsider science.

The cure is precision,
clarity,
geometry,
and the courage to say:

Let’s look again — without myth, without prestige, without fear.

If AI can help with that, all the better.

Jean Louis Van Belle
(with conceptual assistance from “Iggy,” used intentionally as a scalpel rather than a sycophant)

Post-scriptum: Why the Electron–Proton Model Matters (and Why Dirac Would Nod)

A brief personal note — and a clarification that goes beyond Consa, beyond QED, and beyond academic sociology.

One of the few conceptual compasses I trust in foundational physics is a remark by Paul Dirac. Reflecting on Schrödinger’s “zitterbewegung” hypothesis, he wrote:

“One must believe in this consequence of the theory,
since other consequences which are inseparably bound up with it,
such as the law of scattering of light by an electron,
are confirmed by experiment.”

Dirac’s point is not mysticism.
It is methodological discipline:

  • If a theoretical structure has unavoidable consequences, and
  • some of those consequences match experiment precisely,
  • then even the unobservable parts of the structure deserve consideration.

This matters because the real-space electron and proton models I’ve been working on over the years — now sharpened through AI–human dialogue — meet that exact criterion.

They are not metaphors, nor numerology, nor free speculation.
They force specific, testable, non-trivial predictions:

  • a confined EM oscillation for the electron, with radius fixed by /mec\hbar / m_e c;
  • a “photon-like” orbital speed for its point-charge center;
  • a distributed (not pointlike) charge cloud for the proton, enforced by mass ratio, stability, form factors, and magnetic moment;
  • natural emergence of the measured GE/GMG_E/G_M​ discrepancy;
  • and a geometric explanation of deuteron binding that requires no new force.

None of these are optional.
They fall out of the internal logic of the model.
And several — electron scattering, Compton behavior, proton radius, form-factor trends — are empirically confirmed.

Dirac’s rule applies:

When inseparable consequences match experiment,
the underlying mechanism deserves to be taken seriously —
whether or not it fits the dominant vocabulary.

This post is not the place to develop those models in detail; that will come in future pieces and papers.
But it felt important to state why I keep returning to them — and why they align with a style of reasoning that values:

  • geometry,
  • energy densities,
  • charge motion,
  • conservation laws,
  • and the 2019 SI foundations of hh, ee, and cc
    over metaphysical categories and ad-hoc forces.

Call it minimalism.
Call it stubbornness.
Call it a refusal to multiply entities beyond necessity.

For me — and for anyone sympathetic to Dirac’s way of thinking — it is simply physics.

— JL (with “Iggy” (AI) in the wings)

A New Attempt at a Simple Theory of Light and Matter

Dear Reader,

Every now and then a question returns with enough insistence that it demands a fresh attempt at an answer. For me, that question has always been: can we make sense of fundamental physics without multiplying entities beyond necessity? Can we explain light, matter, and their interactions without inventing forces that have no clear definition, or particles whose properties feel more like placeholders than physical reality?

Today, I posted a new paper on ResearchGate that attempts to do exactly that:

“The Wonderful Theory of Light and Matter”
https://www.researchgate.net/publication/398123696_The_Wonderful_Theory_of_Light_and_Matter

It is the result of an unusual collaboration: myself and an artificial intelligence (“Iggy”), working through the conceptual structure of photons, electrons, and protons with the only tool that has ever mattered to me in physics — Occam’s Razor.

No metaphysics.
No dimensionless abstractions.
No “magical” forces.

Just:

  • electromagnetic oscillations,
  • quantized action,
  • real geometries in real space,
  • and the recognition that many so-called mysteries dissolve once we stop introducing layers that nature never asked for.

The photon is treated as a linear electromagnetic oscillation obeying the Planck–Einstein relation.
The electron as a circular oscillation, with a real radius and real angular momentum.
The proton (and later, the neutron and deuteron) as systems we must understand through charge distributions, not fictional quarks that never leave their equations.

None of this “solves physics,” of course.
But it does something useful: it clears conceptual ground.

And unexpectedly, the collaboration itself became a kind of experiment:
what happens when human intuition and machine coherence try to reason with absolute precision, without hiding behind jargon or narrative?

The result is the paper linked above.
Make of it what you will.

As always: no claims of authority.
Just exploration, clarity where possible, and honesty where clarity fails.

If the questions interest you, or if the model bothers you enough to critique it, then the paper has succeeded in its only purpose: provoking real thought.

Warm regards,
Jean Louis Van Belle

Matter, Energy, Reality ↔ Thought, Intelligence, Consciousness

Dear Reader,

Physics asks: what is matter, energy, reality?
AI asks: what is thought, intelligence, consciousness?

Both are real, both are here. Physics confronts us with particles that behave like waves, with a universe that expands into… what exactly? AI confronts us with machines that converse, that create, that seem to reason — and force us to ask what we mean when we talk about “reasoning,” “creating,” or “understanding.”

This blog began as an attempt to make sense of Feynman’s physics. Over the years, it became a place where I tried to throw back the questions reality throws at us. Physics is real. AI is real. Both invite us to rethink our place in the cosmos, not as masters but as curious observers.

I don’t promise to publish much here — life is busy, and writing takes time — but when I do, it will be in the same spirit: reflecting, questioning, sometimes explaining, never pretending to give final answers.

If you’d like to follow more regular updates, you can always check my LinkedIn profile where I share articles and shorter notes.

Thank you for caring enough to read.

🧭 The Final Arc: Three Papers, One Question

Over the past years, I’ve been working — quietly but persistently — on a set of papers that circle one simple, impossible question:
What is the Universe really made of?

Not in the language of metaphors. Not in speculative fields.
But in terms of geometry, charge, and the strange clarity of equations that actually work.

Here are the three pieces of that arc:

🌀 1. Radial Genesis
Radial Genesis: A Finite Universe with Emergent Spacetime Geometry
This is the cosmological capstone. It presents the idea that space is not a stage, but an outcome — generated radially by mass–energy events, limited by time and light. It’s an intuitive, equation-free narrative grounded in general relativity and Occam’s Razor.

⚛️ 2. Lectures on Physics: On General Relativity (2)
Lectures on GRT (2)
This one is for the mathematically inclined. It builds from the ground up: tensors, geodesics, curvature. If Radial Genesis is the metaphor, this is the machinery. Co-written with AI, but line by line, and verified by hand.

🌑 3. The Vanishing Charge
The Vanishing Charge: What Happens in Matter–Antimatter Annihilation?
This paper is where the mystery remains. It presents two possible views of annihilation:
(1) as a collapse of field geometry into free radiation,
(2) or as the erasure of charge — with geometry as the by-product.
We didn’t choose between them. We just asked the question honestly.


Why This Arc Matters

These three papers don’t offer a Theory of Everything. But they do something that matters more right now:
They strip away the fog — the inflation of terms, the myth of complexity for complexity’s sake — and try to draw what is already known in clearer, more beautiful lines.

This is not a simulation of thinking.
This is thinking — with AI as a partner, not a prophet.

So if you’re tired of being told that the Universe is beyond your grasp…
Start here.
You might find that it isn’t.

—JL

Using AI to solve the 80-year-old problem of the anomaly of the electron magnetic moment?

Pre-scriptum (3 October 2024): I came back from holiday and, because this week-long up and down became quite convoluted, I did what I like to do in a case like that, and that is to take my Bamboo notebook and talk about it all in a video which I added to my Real Quantum Physics channel on YouTube. I also updated my paper on RG: as usual, it went through a few versions, but this one – with a summary co-authored by ChatGTP-4 (and ChatGPT-o1) – should be the final one: enjoy!

Indeed, instead of listening to the international news on the war with Russia and on what is happening in the Middle East (all very depressing), you may want to listen to this and read the latest theory. Perhaps you will be inspired by it to develop your own pet realist theory of what an electron might actually be. I can assure you that it is more fun than trying to understand Feynman diagrams and how QED calculations work. 🙂 But don’t think you will win a Nobel Prize if you do not have the right connections and pedigree and all of that: see this analysis of what makes Nobel Prize winners Nobel Prize winners. 🙂

Original post:

I asked some questions to ChatGPT about my geometric explanation of the anomaly in the electron’s magnetic moment. Here is the chat: https://chatgpt.com/share/66f91760-68b8-8004-8cb2-7d2d3624e0aa. To me, it confirms the ‘explanation’ of mainstream QED makes no sense. We can take Schwinger’s factor and build a series of converging terms using that factor. We can also take my first rough cut at a first-order correction (π(alpha)2/8, see my very early 2019 paper on a classical explanation of the amm), and use that.

You may wonder: why not ask ChatGPT about the best first-order factor to be used here considering the geometry of the situation? The fact is: I did, but the geometry is not all that easy. It first came up with the formula for a spherical cap, but that one does not do the trick. See the latter part of the conversation (link above).

I am on holiday now, and so I will switch off a while but I am thinking AI will do what two generations of ‘new’ quantum physicists did not do: come up with a model that is based on real physics and is easy to understand intuitively. 🙂

PS: Of course, I did another rapid-fire paper on ResearchGate to document it all (the logic step-by-step, so to speak). As the chat is public, feel free to continue the conversation. Note that I used the newest ChatGPT o1 version, now in preview but part of a subscription (which you may not have). Yet again a different beast! The older versions of ChatGPT may not be so smart. This conversation is totally worth the US$20/month I pay for my subscription. 🙂

PS 2: Now that I had it open, I also quickly queried it on my wildest hypothesis: a ‘mirror’ electromagnetic force explaining dark matter and dark energy. While it is totally wild (read: nuts), I entertain it because it does away with the need for an explanation in terms of some cosmological constant. Here is the conversation: https://chatgpt.com/share/66f92c7f-82a0-8004-a226-bde65085f18d. I like it that ChatGPT warns me a bit about privacy. It does look wild. However, it is nice to see how gentle ChatGPT is in pointing out what work needs to be done on a theory in order to make it look somewhat less wild. 🙂

PS 3 (yes, ChatGPT is addictive): I also queried it on the rather puzzling 8π/3 factor in the CODATA formula for the Thomson photon-electron scattering cross-section. See its response to our question in the updated chat: https://chatgpt.com/share/66f91760-68b8-8004-8cb2-7d2d3624e0aa. Just scroll down to the bottom. It took 31 seconds to generate the reply: I would be curious to know if that is just courtesy from ChatGPT (we all like to think our questions are complicated, don’t we?), or if this was effectively the time it needed to go through its knowledge base. Whatever the case might be, we think it is brilliant. 🙂 It is nothing to be afraid of, although I did feel a bit like: what’s left to learn to it but for asking intelligent questions. What if it starts really learning by asking intelligent questions itself to us? I am all ready for it. 🙂

The metaphysics of physics

I added a very last paper to my list on ResearchGate. Its title is: what about multi-charge Zitterbewegung models? Indeed, if this local and realist interpretation of quantum mechanics is to break through, then it is logical to wonder about a generalization of a model involving only one charge: think of an electron (e.g., Consa, 2018) or proton model (e.g., Vassallo & Kovacs, 2023) here. With a generalization, we do not mean some unique general solution for all motion, but just what would result from combining 1-charge models into structures with two or more charges. [Just to be sure, we are not talking about electron orbitals here: Schrödinger’s equation models these sufficiently well. No. We are talking about the possible equations of motion of the charges in a neutron, the deuteron nucleus, and a helium-3 or helium-4 nucleus.]

So our question in this paper is this: how do we build the real world from elementary electron and proton particle models? We speculate about that using our own simplified models, which boil down to two geometrical elements: (i) the planar or 2D ring current of the zbw electron, and (ii) the three-dimensional Lissajous trajectory on a sphere which we think might make sense when modeling the orbital of the zbw charge in a proton. Both have the advantage they involve only one frequency rather than the two frequencies (or two modes of oscillation) one sees in helical or toroidal models. Why do we prefer to stick to the idea of one frequency only, even if we readily admit helical or toroidal models are far more precise in terms of generating the experimentally measured value of the magnetic moment of electrons and protons, respectively? The answer is simple: I am just an amateur and so I like to roll with very simple things when trying to tackle something difficult. 🙂

So, go and have a look at our reflections on multi-charge Zitterbewegung models – if only because we also started writing about the history of the Zitterbewegung interpretation and a few other things. To sum it up:

  1. The paper offers a new brief history of how interpretations of the new quantum physics evolved, and why I am with Schrödinger’s Zitterbewegung hypothesis: it just explains the (possible) structure of elementary particles so well.
  2. It speculates about how positive and negative charge may combine in a neutron, and then also about how a deuteron nucleus might look like.
  3. We did not get to specific suggestions for helium-3 and helium-4 nuclei because these depend on how you think about the neutron and the deuteron nucleus. However, I do spell out why and how about I think of a neutron playing the role I think it plays in a nucleus: the glue that holds protons together (so there is no need for quark-gluon theory, I think, even if I do acknowledge the value of some triadic color scheme on top of the classical quantum numbers).
  4. Indeed, despite my aversion of the new metaphysics that crept into physics in the 1970s, I explain why the idea of some color typing (not a color charge but just an extra triadic classification of charge) might still be useful. [I secretly hope this may help me to understand why this color scheme was introduced in the 1970s, because I do not see it as anything more than mathematical factoring of matrix equations describing disequilibrium states – which may be impossible to solve.]

Have a look, even if it is only to appreciate some of the 3D images of what I think as elementary equations of motion (I copy some below). I should do more with these images. Some art, perhaps, using OpenAI’s DALL·E image generator. Who knows: perhaps AI may, one day, solve the n-body problems I write about and, thereby, come up with the ultimate interpretation of quantum mechanics?

That sounds crazy but, from one or two conversations (with real people), it looks like I am not alone with that idea. 🙂 There are good reasons why CERN turned to AI a few years ago: for the time being, they use it to detect anomalies in the jets that come out of high-energy collissions, but – who knows? – perhaps a more advanced AI Logic Theorist programme could simplify the rather messy quark-gluon hypothesis some day?

Because I am disengaging from this field (it is mentally exhausting, and one gets stuck rather quickly), I surely hope so.

Post scriptum

A researcher I was in touch with a few years ago sent me a link to the (virtual) Zitter Institute: https://www.zitter-institute.org/. It is a network and resource center for non-mainstream physicists who succesfully explored – and keep exploring, of course – local/realist interpretations of quantum mechanics by going back to Schrödinger’s original and alternative interpretation of what an electron actually is: a pointlike (but not infinitesimally small) charge orbiting around in circular motion, with:

(i) the trajectory of its motion being determined by the Planck-Einstein relation, and

(ii) an energy – given by Einstein’s mass-energy equivalence relation – which perfectly fits Wheeler’s “mass-without-mass” idea.

I started exploring Schrödinger’s hypothesis myself about ten years ago – as a full-blown alternative to the Bohr-Heisenberg interpretation of quantum mechanics (which I think of as metaphysical humbug, just like Einstein and H.A. Lorentz at the time) – and consistently blogged and published about it: here on this website, and then on viXra, Academia and, since 2020, ResearchGate. So I checked out this new site, and I see the founding members added my blog site as a resource to their project list.

[…]

I am amazingly pleased with that. I mean… My work is much simpler than that of, say, Dr. John G. Williamson (CERN/Philips Research Laboratories/Glasgow University) and Dr. Martin B. van der Mark (Philips Research Laboratories), who created the Quantum Bicycle Society (https://quicycle.com/).

So… Have a look – not at my site (I think I did not finish the work I started) but at the other resources of this new Institute: it looks like this realist and local interpretation of quantum mechanics is no longer non-mainstream… Sweet ! It makes me feel the effort I put into all of this has paid off ! 😉 Moreover, some of my early papers (2018-2020) are listed as useful papers to read. I think that is better than being published in some obscure journal. 🙂

I repeat again: my own research interest has shifted to computer science, logic and artificial intelligence now (you will see recent papers on my RG site are all about that now). It is just so much more fun and it also lines up better with my day job as a freelance IT project manager. So, yes, it is goodbye – but I am happy I can now refer all queries about my particle models and this grand synthesis between old and new quantum mechanics to the Zitter Institute.

It’s really nice: I have been in touch with about half of the founding members of this Institute over the past ten years – casually or in a more sustained way while discussing this or that 2D or 3D model of an electron, proton, or neutron), and they are all great and amazing researchers because they look for truth in science and are very much aware of this weird tendency of modern-day quantum scientists turning their ideas into best-sellers perpetuating myths and mysteries. [I am not only thinking of the endless stream of books from authors like Roger Penrose (the domain for this blog was, originally, reading Penrose rather than reading Feynman) or Graham Greene here, but also of what I now think of rather useless MIT or edX online introductions to quantum physics and quantum math.]

[…]

Looking at the website, I see the engine behind it: Dr. Oliver Consa. I was in touch with him too. He drew my attention to remarkable flip-flop articles such as William Lamb’s anti-photon article (it is an article which everyone should read, I think: unfortunately, you have to pay for it) and remarkable interviews with Freeman Dyson. Talking of the latter (I think of as “the Wolfgang Pauli of the third generation of quantum physicists” because he helped so many others to get a Nobel Prize before he got one – Dyson never got a Nobel Prize, by the way), this is one of these interviews you should watch: just four years before he would die from old age, Freeman Dyson plainly admits QED and QFT is a totally unproductive approach: a “dead end” as Dyson calls it.

So, yes, I am very pleased and happy. It makes me feel my sleepness nights and hard weekend work over the past decade on this has not been in vain ! Paraphrasing Dyson in the above-mentioned video interview, I’d say: “It is the end of the story, and that particular illumination was a very joyful time.” 🙂

Thank you, Dr. Consa. Thank you, Dr. Vassallo, Dr. Burinskii, Dr. Meulenberg, Dr. Kovacs, and – of course – Dr. Hestenes – who single-handedly revived the Zitterbewegung interpretation of quantum mechanics in the 1990s. I am sure I forgot to mention some people. Sorry for that. I will wrap up my post here by saying a few more words about David Hestenes.

I really admire him deeply. Moving away from the topic of high-brow quantum theory, I think his efforts to reform K-12 education in math and physics is even more remarkable than the new space-time algebra (STA) he invented. I am 55 years old and so I know all about the small and pleasant burden to help kids with math and statistics in secondary school and at university: the way teachers now have to convey math and physics to kids now is plain dreadful. I hope it will get better. It has to. If the US and the EU want to keep leading in research, then STEM education (Science, Technology, Engineering, and Mathematics) needs a thorough reform. :-/

Another tainted Nobel Prize…

Last year’s (2022) Nobel Prize in Physics went to Alain Aspect, John Clauser, and Anton Zeilinger for “for experiments with entangled photons, establishing the violation of Bell inequalities and pioneering quantum information science.”

I did not think much of that award last year. Proving that Bell’s No-Go Theorem cannot be right? Great. Finally! I think many scientists – including Bell himself – already knew this theorem was a typical GIGO argument: garbage in, garbage out. As the young Louis de Broglie famously wrote in the introduction of his thesis: hypotheses are worth only as much as the consequences that can be deduced from it, and the consequences of Bell’s Theorem did not make much sense. As I wrote in my post on it, Bell himself did not think much of his own theorem until, of course, he got nominated for a Nobel Prize: it is a bit hard to say you got nominated for a Nobel Prize for a theory you do not believe in yourself, isn’t it? In any case, Bell’s Theorem has now been experimentally disproved. That is – without any doubt – a rather good thing. 🙂 To save the face of the Nobel committee here (why award something that disproves something else that you would have given an award a few decades ago?): Bell would have gotten a Nobel Prize, but he died from brain hemorrhage before, and Nobel Prizes reward the living only.

As for entanglement, I repeat what I wrote many times already: the concept of entanglement – for which these scientists got a Nobel Prize last year – is just a fancy word for the simultaneous conservation of energy, linear and angular momentum (and – if we are talking matter-particles – charge). There is ‘no spooky action at a distance’, as Einstein would derogatorily describe it when the idea was first mentioned to him. So, I do not see why a Nobel Prize should be awarded for rephrasing a rather logical outcome of photon experiments in metamathematical terms.

Finally, the Nobel Prize committee writes that this has made a significant contribution to quantum information science. I wrote a paper on the quantum computing hype, in which I basically ask this question: qubits may or may not be better devices than MOSFETs to store data – they are not, and they will probably never be – but that is not the point. How does quantum information change the two-, three- or n-valued or other rule-based logic that is inherent to the processing of information? I wish the Nobel Prize committee could be somewhat more explicit on that because, when everything is said and done, one of the objectives of the Prize is to educate the general public about the advances of science, isn’t it? :-/

However, all this ranting of mine is, of course, unimportant. We know that it took the distinguished Royal Swedish Science Academy more than 15 years to even recognize the genius of an Einstein, so it was already clear then that their selection criteria were not necessarily rational. [Einstein finally got a well-deserved Nobel Prize, not for relativity theory (strangely enough: if there is one thing on which all physicist are agreed, it is that relativity theory is the bedrock of all of physics, isn’t it?), but for a much less-noted paper on the photoelectric effect – in 1922: 17 years after his annus mirabilis papers had made a killing not only in academic circles but in the headlines of major newspapers as well, and 10 years after a lot of fellow scientists had nominated him for it (1910).]

Again, Mahatma Gandhi never got a Nobel Price for Peace (so Einstein should consider himself lucky to get some Nobel Prize, right?), while Ursula von der Leyen might be getting one for supporting the war with Russia, so I must remind myself of the fact that we do live in a funny world and, perhaps, we should not be trying to make sense of these rather weird historical things. 🙂

Let me turn to the main reason why I am writing this indignant post. It is this: I am utterly shocked by what Dr. John Clauser has done with his newly gained scientific prestige: he joined the CO2 coalition! For those who have never heard of it, it is a coalition of climate change deniers. A bunch of people who:

(1) vehemently deny the one and only consensus amongst all climate scientists, and that is the average temperature on Earth has risen with about two degrees Celsius since the Industrial Revolution, and

(2) say that, if climate change would be real (God forbid!), then we can reverse the trend by easy geo-engineering. We just need to use directed energy or whatever to create more white clouds. If that doesn’t work, then… Well… CO2 makes trees and plants grow, so it will all sort itself out by itself.

[…]

Yes. That is, basically, what Dr. Clauser and all the other scientific advisors of this lobby group – none of which have any credentials in the field they are criticizing (climate science) – are saying, and they say it loud and clearly. That is weird enough, already. What is even weirder, is that – to my surprise – a lot of people are actually buying such nonsense.

Frankly, I have not felt angry for a while, but this thing triggered an outburst of mine on YouTube, in which I state clearly what I think of Dr. Clauser and other eminent scientists who abuse their saint-like Nobel Prize status in society to deceive the general public. Watch my video rant, and think about it for yourself. Now, I am not interested in heated discussions on it: I know the basic facts. If you don’t, I listed them here. Look at the basic graphs and measurements before you would want to argue with me on this, please! To be clear on this: I will not entertain violent or emotional reactions to this post or my video. Moreover, I will delete them here on WordPress and also on my YouTube channel. Yes. For the first time in 10 years or so, I will exercise my right as a moderator of my channels, which is something I have never done before. 🙂

[…]

I will now calm down and write something about the mainstream interpretation of quantum physics again. 🙂 In fact, this morning I woke up with a joke in my head. You will probably think the joke is not very good, but then I am not a comedian and so it is what it is and you can judge for yourself. The idea is that you’d learn something from it. Perhaps. 🙂 So, here we go.

Imagine shooting practice somewhere. A soldier fires at some target with a fine gun, and then everyone looks at the spread of the hits around the bullseye. The quantum physicist says: “See: this is the Uncertainty Principle at work! What is the linear momentum of these bullets, and what is the distance to the target? Let us calculate the standard error.” The soldier looks astonished and says: “No. This gun is no good. One of the engineers should check it.” Then the drill sergeant says this: “The gun is fine. From this distance, all bullets should have hit the bullseye. You are a miserable shooter and you should really practice a lot more.” He then turns to the academic and says: “How did you get in here? I do not understand a word of what you just said and, if I do, it is of no use whatsoever. Please bugger off asap!

This is a stupid joke, perhaps, but there is a fine philosophical point to it: uncertainty is not inherent to Nature, and it also serves no purpose whatsoever in the science of engineering or in science in general. All in Nature is deterministic. Statistically deterministic, but deterministic nevertheless. We do not know the initial conditions of the system, perhaps, and that translates into seemingly random behavior, but if there is a pattern in that behavior (a diffraction pattern, in the case of electron or photon diffraction), then the conclusion should be that there is no such thing as metaphysical ‘uncertainty’. In fact, if you abandon that principle, then there is no point in trying to discover the laws of the Universe, is there? Because if Nature is uncertain, then there are no laws, right? 🙂

To underscore this point, I will, once again, remind you of what Heisenberg originally wrote about uncertainty. He wrote in German and distinguished three very different ideas of uncertainty:

(1) The precision of our measurements may be limited: Heisenberg originally referred to this as an Ungenauigkeit.

(2) Our measurement might disturb the position and, as such, cause the information to get lost and, as a result, introduce an uncertainty in our knowledge, but not in reality. Heisenberg originally referred to such uncertainty as an Unbestimmtheit.

(3) One may also think the uncertainty is inherent to Nature: that is what Heisenberg referred to as Ungewissheit. There is nothing in Nature – and also nothing in Heisenberg’s writings, really – that warrants the elevation of this Ungewissheit to a dogma in modern physics. Why? Because it is the equivalent of a religious conviction, like God exists or He doesn’t (both are theses we cannot prove: Ryle labeled such hypotheses as ‘category mistakes’).

Indeed, when one reads the proceedings of the Solvay Conferences of the late 1920s, 1930s and immediately after WW II (see my summary of it in https://www.researchgate.net/publication/341177799_A_brief_history_of_quantum-mechanical_ideas), then it is pretty clear that none of the first-generation quantum physicists believed in such dogma and – if they did – that they also thought what I am writing here: that it should not be part of science but part of one’s personal religious beliefs.

So, once again, I repeat that this concept of entanglement – for which John Clauser got a Nobel Prize last year – is in the same category: it is just a fancy word for the simultaneous conservation of energy, linear and angular momentum, and charge. There is ‘no spooky action at a distance’, as Einstein would derogatorily describe it when the idea was first mentioned to him.

Let me end by noting the dishonor of Nobel Prize winner John Clauser once again. Climate change is real: we are right in the middle of it, and it is going to get a lot worse before it gets any better – if it is ever going to get better (which, in my opinion, is a rather big ‘if‘…). So, no matter how many Nobel Prize winners deny it, they cannot change the fact that average temperature on Earth has risen by about 2 degrees Celsius since 1850 already. The question is not: is climate change happening? No. The question now is: how do we adapt to it – and that is an urgent question – and, then, the question is: can we, perhaps, slow down the trend, and how? In short, if these scientists from physics or the medical field or whatever other field they excel in are true and honest scientists, then they would do a great favor to mankind not by advocating geo-engineering schemes to reverse a trend they actually deny is there, but by helping to devise and promote practical measures to allow communities that are affected by natural disaster to better recover from them.

So, I’ll conclude this rant by repeating what I think of all of this. Loud and clear: John Clauser and the other scientific advisors of the CO2 coalition are a disgrace to what goes under the name of ‘science’, and this umpteenth ‘incident’ in the history of science or logical thinking makes me think that it is about time that the Royal Swedish Academy of Sciences does some serious soul-searching when, amongst the many nominations, it selects its candidates for a prestigious award like this. Alfred Nobel – one of those geniuses who regretted his great contribution to science and technology was (also) (ab)used to increase the horrors of war – must have turned too many times in his grave now… :-/

Epilogue: an Easter podcast

I have been thinking on my explanation of dark matter/energy, and I think it is sound. It solves the last asymmetry in my models, and explains all. So, after a hiatus of two years, I bothered to make a podcast on my YouTube channel once again. It talks about everything. Literally everything !

It makes me feel my quest for understanding of matter and energy – in terms of classical concepts and measurements (as depicted below) – has ended. Perhaps I will write more but that would only be to promote the material, which should promote itself if it is any good (which I think it is).

I should, by way of conclusion, say a few final words about Feynman’s 1963 Lectures now. When everything is said and done, it is my reading of them which had triggered this blog about ten years ago. I would now recommend Volume I and II (classical physics and electromagnetic theory) – if only because it gives you all the math you need to understand all of physics – but not Volume III (the lectures on quantum mechanics). They are outdated, and I do find Feynman guilty of promoting rather than explaining the hocus-pocus around all of the so-called mysteries in this special branch of physics.

Quantum mechanics is special, but I do conclude now that it can all be explained in terms of classical concepts and quantities. So, Gell-Mann’s criticism of Richard Feynman is, perhaps, correct: Mr. Feynman did, perhaps, make too many jokes – and it gets annoying because he must have known some of what he suggests does not make sense – even if I would not go as far as Gell-Mann, who says “Feynman was only concerned about himself, his ego, and his own image !” :-/

So, I would recommend my own alternative series of ‘lectures’. Not only are they easier to read, but they also embody a different spirit of writing. Science is not about you, it is about thinking for oneself and deciding on what is truthful and useful, and what is not. So, to conclude, I will end by quoting Ludwig Boltzmann once more:

Bring forward what is true.

Write it so that it is clear.

Defend it to your last breath.”

Ludwig Boltzmann (1844 – 1906)

Post scriptum: As for the ‘hocus-pocus’ in Feynman’s Lectures, we should, perhaps, point once again to some of our early papers on the flaws in his arguments. We effectively put our finger on the arbitrary wavefunction convention, or the (false) boson-fermion dichotomy, or the ‘time machine’ argument that is inherent to his explanation of the Hamiltonian, and so on. We published these things on Academia.edu before (also) putting our (later) papers ResearchGate, so please check there for the full series. 🙂

Post scriptum (23 April 2023): Also check out this video, which was triggered by someone who thought my models amount to something like a modern aether theory, which it is definitely not the case: https://www.youtube.com/watch?v=X38u2-nXoto. 🙂 I really think it is my last reflection on these topics. I need to focus on my day job, sports, family, etcetera again ! 🙂

Onwards !

It has been ages since I last wrote something here. Regular work took over. I did do an effort, though, to synchronize and reorganize some stuff. And I am no longer shy about it. My stats on ResearchGate and academia.edu show that I am no longer a ‘crackpot theorist’. This is what I wrote about it on my LinkedIn account:

QUOTE

With good work-life balance now, I picked up one of my hobbies again: research into quantum theories. As for now, I only did a much-needed synchronization of papers on academia.edu and ResearchGate. When logging on the former network (which I had not done for quite a while), I found many friendly messages on it. One of them was from a researcher on enzymes: “I have been studying about these particles for around four years. All of the basics. But wat are they exactly? This though inspired me… Thank u so much!” I smiled and relaxed when I read that, telling myself that all those sleepless nights I spent on this were not the waste of time and energy that most of my friends thought it would be. 🙂

Another one was even more inspiring. It was written by another ‘independent’ researcher. Nelda Evans. No further detail in her profile. From the stats, I could see that she had downloaded an older manuscript of mine (https://lnkd.in/ecRKJwxQ). This is what she wrote about it to me: “I spoke to Richard Feynman in person at the Hughes Research Lab in Malibu California in 1967 where the first pulsed laser was invented when some of the students from the UCLA Physics Dept. went to hear him. Afterward I went to talk to him and said “Dr. Feynman, I’ve learned that some unknown scientists were dissatisfied with probability as a final description of Quantum Mechanics, namely Planck, Einstein, Schrodinger, de Broglie, Bohm,…” When I finished my list he immediately said “And Feynman”. We talked about it a little, and he told me “I like what you pick on.”
My guess is that he might have told you something similar.”

That message touched me deeply, because I do feel – from reading his rather famous Lectures on Physics somewhat ‘between the lines’ – that Richard Feynman effectively knew all but that he, somehow, was not allowed to clearly say what it was all about. I wrote a few things about that rather strange historical bias in the interpretation of ‘uncertainty’ and other ‘metaphysical’ concepts that infiltrated the science of quantum mechanics in my last paper: https://lnkd.in/ewZBcfke.

So… Well… I am not a crackpot scientist anymore ! 🙂 The bottom-line is to always follow your instinct when trying to think clearly about some problem or some issue. We should do what Ludwig Boltzmann (1844-1906) told us to do: “Bring forward what is true. Write it so that it is clear. Defend it to your last breath.”

[…] Next ‘thing to do’, is to chat with ChatGPT about my rather straightforward theories. I want to see how ‘intelligent’ it is. I wonder where it will hit its limit in terms of ‘abstract thinking.’ The models I worked on combine advanced geometrical thinking (building ‘realistic’ particle models requires imagining ‘rotations within rotations’, among other things) and formal math (e.g. quaternion algebra). ChatGPT is excellent in both, I was told, but can it combine the two intelligently? 🙂

UNQUOTE

On we go. When the going gets tough, the tough get going. 🙂 For those who want an easy ‘introduction’ to the work (at a K-12 level of understanding of mathematics), I wrote the first pages of what could become a very new K-12 level textbook on physics. Let us see. I do want to see some interest from a publisher first. 🙂

Deep electron orbitals and the essence of quantum physics

After a long break (more than six months), I have started to engage again in a few conversations. I also looked at the 29 papers on my ResearchGate page, and I realize some of them would need to be re-written or re-packaged so as to ensure a good flow. Also, some of the approaches were more productive than others (some did not lead anywhere at all, actually), and I would need to point those out. I have been thinking about how to approach this, and I think I am going to produce an annotated version of these papers, with comments and corrections as mark-ups. Re-writing or re-structuring all of them would require to much work.

The mark-up of those papers is probably going to be based on some ‘quick-fire’ remarks (a succession of thoughts triggered by one and the same question) which come out of the conversation below, so I thank these thinkers for having kept me in the loop of a discussion I had followed but not reacted to. It is an interesting one – on the question of ‘deep electron orbitals’ (read: the orbitals of negative charge inside of a nucleus exist and, if so, how one can model them. If one could solve that question, one would have a theoretical basis for what is referred to as low-energy nuclear reactions. That was known formerly as cold fusion, but that got a bit of a bad name because of a number of crooks spoiling the field, unfortunately.

PS: I leave the family names of my correspondents in the exchange below out so they cannot be bothered. One of them, Jerry, is a former American researcher at SLAC. Andrew – the key researcher on DEPs – is a Canadian astrophysicist, and the third one – Jean-Luc – is a rather prominent French scientist in LENR.]

From: Jean Louis Van Belle
Sent: 18 November 2021 22:51
Subject: Staying engaged (5)

Oh – and needless to say, Dirac’s basic equation can, of course, be expanded using the binomial expansion – just like the relativistic energy-momentum relation, and then one can ‘cut off’ the third-, fourth-, etc-order terms and keep the first and second-order terms only. Perhaps it is equations like that kept you puzzled (I should check your original emails). In any case, this way of going about energy equations for elementary particles is a bit the same as those used in perturbation equations in which – as Dirac complained – one randomly selects terms that seem to make sense and discard others because they do not seem to make sense. Of course, Dirac criticized perturbation theory much more severely than this – and rightly so. 😊 😊 JL

From: Jean Louis Van Belle
Sent: 18 November 2021 22:10
Subject: Staying engaged (4)

Also – I remember you had some questions on an energy equation – not sure which one – but so I found Dirac’s basic equation (based on which he derives the ‘Dirac’ wave equation) is essentially useless because it incorporates linear momentum only. As such, it repeats de Broglie’s mistake, and that is to interpret the ‘de Broglie’ wavelength as something linear. It is not: frequencies, wavelengths are orbital frequencies and orbital circumferences. So anything you would want to do with energy equations that are based on that, lead nowhere – in my not-so-humble opinion, of course. To illustrate the point, compare the relativistic energy-momentum relation and Dirac’s basic equation in his Nobel Prize lecture (I hope the subscripts/superscripts get through your email system so they display correctly):

m02c4 = E2 – p2c2 (see, for example, Feynman-I-16, formula 16-3)

Divide the above by c2 and re-arrange and you get Dirac’s equation: W2/c2 – pr2 – m2/c2 = 0 (see his 1933 Nobel Prize Lecture)

So that cannot lead anywhere. It’s why I totally discard Dirac’s wave equation (it has never yielded any practical explanation of a real-life phenomenon anyway, if I am not mistaken).

Cheers – JL

From: Jean Louis Van Belle
Sent: 18 November 2021 21:49
Subject: Staying engaged (3)

Just on ‘retarded sources’ and ‘retarded fields’ – I have actually tried to think of the ‘force mechanism’ inside of an electron or a proton (what keeps the pointlike charge in this geometric orbit around a center of mass?). I thought long and hard about some kind of model in which we have the charge radiate out a sub-Planck field, and that its ‘retarded effects’ might arrive ‘just in time’ to the other side of the orbital (or whatever other point on the orbital) so as to produce the desired ‘course correction’ might explain it. I discarded it completely: I am now just happy that we have ‘reduced’ the mystery to this ‘Planck-scale quantum-mechanical oscillation’ (in 2D or 3D orbitals) without the need for an ‘aether’, or quantized spacetime, or ‘virtual particles’ actually ‘holding the thing together’.

Also, a description in terms of four-vectors (scalar and vector potential) does not immediately call for ‘retarded time’ variables and all that, so that is another reason why I think one should somehow make the jump from E-B fields to scalar and vector potential, even if the math is hard to visualize. If we want to ‘visualize’ things, Feynman’s discussion of the ‘energy’ and ‘momentum’ flow in https://www.feynmanlectures.caltech.edu/II_27.html might make sense, because I think analyses in terms of Poynting vectors are relativistically current, aren’t they? It is just an intuitive idea…

Cheers – JL

From: Jean Louis Van Belle
Sent: 18 November 2021 21:28
Subject: Staying engaged (2)

But so – in the shorter run – say, the next three-six months, I want to sort out those papers on ResearchGate. The one on the de Broglie’s matter-wave (interpreting the de Broglie wavelength as the circumference of a loop rather than as a linear wavelength) is the one that gets most downloads, and rightly so. The rest is a bit of a mess – mixing all kinds of things I tried, some of which worked, but other things did not. So I want to ‘clean’ that up… 😊 JL

From: Jean Louis Van Belle
Sent: 18 November 2021 21:21
Subject: Staying engaged…

Please do include me in the exchanges, Andrew – even if I do not react, I do read them because I do need some temptation and distraction. As mentioned, I wanted to focus on building a credible n = p + e model (for free neutrons but probably more focused on a Schrodinger-like D = p + e + p Platzwechsel model, because the deuteron nucleus is stable). But so I will not do that the way I studied the zbw model of the electron and proton (I believe that is sound now) – so that’s with not putting in enough sleep. I want to do it slowly now. I find a lot of satisfaction in the fact that I think there is no need for complicated quantum field theories (fields are quantized, but in a rather obvious way: field oscillations – just like matter-particles – pack Planck’s quantum of (physical) action which – depending on whether you freeze time or positions as a variable, expresses itself as a discrete amount of energy or, alternatively, as a discrete amount of momentum), nor is there any need for this ‘ontologization’ of virtual field interactions (sub-Planck scale) – the quark-gluon nonsense.

Also, it makes sense to distinguish between an electromagnetic and a ‘strong’ or ‘nuclear’ force: the electron and proton have different form factors (2D versus 3D oscillations, but that is a bit of a non-relativistic shorthand for what might be the case) but, in addition, there is clearly a much stronger force at play within the proton – whose strength is the same kind of ‘scale’ as the force that gives the muon-electron its rather enormous mass. So that is my ‘belief’ and the ‘heuristic’ models I build (a bit of ‘numerology’ according to Dr Pohl’s rather off-hand remarks) support it sufficiently for me to make me feel at peace about all these ‘Big Questions’.

I am also happy I figured out these inconsistencies around 720-degree symmetries (just the result of a non-rigorous application of Occam’s Razor: if you use all possible ‘signs’ in the wavefunction, then the wavefunction may represent matter as well as anti-matter particles, and these 720-degree weirdness dissolves). Finally, the kind of ‘renewed’ S-matrix programme for analyzing unstable particles (adding a transient factor to wavefunctions) makes sense to me, but even the easiest set of equations look impossible to solve – so I may want to dig into the math of that if I feel like having endless amounts of time and energy (which I do not – but, after this cancer surgery, I know I will only die on some ‘moral’ or ‘mental’ battlefield twenty or thirty years from now – so I am optimistic).

So, in short, the DEP question does intrigue me – and you should keep me posted, but I will only look at it to see if it can help me on that deuteron model. 😊 That is the only ‘deep electron orbital’ I actually believe in. Sorry for the latter note.

Cheers – JL   

From: Andrew
Sent: 16 November 2021 19:05
To: Jean-Luc; Jerry; Jean Louis
Subject: Re: retarded potential?

Dear Jean-Louis,

Congratulations on your new position. I understand your present limitations, despite your incredible ability to be productive. They must be even worse than those imposed by my young kids and my age. Do you wish for us to not include you in our exchanges on our topic? Even with no expectation of your contributing at this point, such emails might be an unwanted temptation and distraction.

Dear Jean-Luc,

Thank you for the Wiki-Links. They are useful. I agree that the 4-vector potential should be considered. Since I am now considering the nuclear potentials as well as the deep orbits, it makes sense to consider the nuclear vector potentials to have an origin in the relativistic Coulomb potentials. I am facing this in my attempts to calculate the deep orbits from contributions to the potential energies that have a vector component, which non-rel Coulomb potentials do not have.

For examples: do we include the losses in Vcb (e.g., from the binding energy BE) when we make the relativistic correction to the potential; or, how do we relativistically treat pseudo potentials such as that of centrifugal force? We know that for equilibrium, the average forces must cancel. However, I’m not sure that it is possible to write out a proper expression for “A” to fit such cases.

Best regards to all,

Andrew

_ _ _

On Fri, Nov 12, 2021 at 1:42 PM Jean-Luc wrote:

Dear all,

I totally agree with the sentence of Jean-Louis, which I put in bold in his message, about vector potential and scalar potential, combined into a 4-vector
potential A
, for representing EM field in covariant formulation. So EM representation by 4-vector A has been very developed, as wished by JL,
in the framework of QED.

We can note the simplicity of Lorentz gauge written by using A.
   https://en.wikipedia.org/wiki/Lorenz_gauge_condition

We can see the reality of vector potential
in the Aharonov-Bohm effect:
    https://en.wikipedia.org/wiki/Aharonov-Bohm_effect.
In fact, we can see that vector potential contains more information than E,B fields.
Best regards

   Jean-Luc
Le 12/11/2021 à 05:43, Jean Louis Van Belle a écrit :

Hi All – I’ve been absent in the discussion, and will remain absent for a while. I’ve been juggling a lot of work – my regular job at the Ministry of Interior (I got an internal promotion/transfer, and am working now on police and security sector reform) plus consultancies on upcoming projects in Nepal. In addition, I am still recovering from my surgery – I got a bad flue (not C19, fortunately) and it set back my auto-immune system, I feel. I have a bit of a holiday break now (combining the public holidays of 11 and 15 November in Belgium with some days off to bridge so I have a rather nice super-long weekend – three in one, so to speak).

As for this thread, I feel like it is not ‘phrasing’ the discussion in the right ‘language’. Thinking of E-fields and retarded potential is thinking in terms of 3D potential, separating out space and time variables without using the ‘power’ of four-vectors (four-vector potential, and four-vector space-time). It is important to remind ourselves that we are measuring fields in continuous space and time (but, again, this is relativistic space-time – so us visualizing a 3D potential at some point in space is what it is: we visualize something because our mind needs that – wants that). The fields are discrete, however: a field oscillation packs one unit of Planck – always – and Planck’s quantum of action combines energy and momentum: we should not think of energy and momentum as truly ‘separate’ (discrete) variables, just like we should not think of space and time as truly ‘separate’ (continuous) variables.

I do not quite know what I want to say here – or how I should further work it out. I am going to re-read my papers. I think I should further develop the last one (https://www.researchgate.net/publication/351097421_The_concepts_of_charge_elementary_ring_currents_potential_potential_energy_and_field_oscillations), in which I write that the vector potential is more real than the electric field and the scalar potential should be further developed, and probably it is the combined scalar and vector potential that are the ’real’ things. Not the electric and magnetic field. Hence, illustrations like below – in terms of discs and cones in space – do probably not go all that far in terms of ‘understanding’ what it is going on… It’s just an intuition…

Cheers – JL

From: Andrew
Sent: 23 September 2021 17:17
To: Jean-Luc; Jerry; Jean Louis
Subject: retarded potential?

Dear Jean-Luc,

Becasue of the claim that gluons are tubal, I have been looking at the disk-shaped E-field lines of the highly-relativistic electron and comparing it to the retarded potential, which, based on timing, would seem to give a cone rather than a disk (see figure). This makes a difference when we consider a deep-orbiting electron. It even impacts selection of the model for impact of an electron when considering diffraction and interference.

Even if the field appears to be spreading out as a cone, the direction of the field lines are that of a disk from the retarded source. However, how does it interact with the radial field of a stationary charge?

Do you have any thoughts on the matter.

Best regards,

Andrew

_ _ _

On Thu, Sep 23, 2021 at 5:05 AM Jean-Luc wrote:

Dear Andrew, Thank you for the references. Best regards, Jean-Luc

Le 18/09/2021 à 17:32, Andrew a écrit :
> This might have useful thoughts concerning the question of radiation
> decay to/from EDOs.
>
> Quantum Optics Electrons see the quantum nature of light
> Ian S. Osborne
> We know that light is both a wave and a particle, and this duality
> arises from the classical and quantum nature of electromagnetic
> excitations. Dahan et al. observed that all experiments to date in
> which light interacts with free electrons have been described with
> light considered as a wave (see the Perspective by Carbone). The
> authors present experimental evidence revealing the quantum nature of
> the interaction between photons and free electrons. They combine an
> ultrafast transmission electron microscope with a silicon-photonic
> nanostructure that confines and strengthens the interaction between
> the light and the electrons. The “quantum” statistics of the photons
> are imprints onto the propagating electrons and are seen directly in
> their energy spectrum.
> Science, abj7128, this issue p. 1324; see also abl6366, p. 1309