Using AI for sense-making once more…

As mentioned in my last post, I did a video (YouTube link here) on why I think the invention of new quantum numbers like strangeness, charm and beauty in the 1960s – and their later ontologization as quarks – makes no sense. As usual, I talk too much and the video is rather long-winding. I asked ChatGPT to make a summary of it, and I think it did a rather good job at that. I copy its summary unaltered below.

Beyond the Quark Hypothesis: A Call for Simplicity in High-Energy Physics

1. Introduction: A Personal Journey in Physics

In this video, I reflect on my path as an amateur physicist reaching 50,000 reads—a milestone that underscores both excitement and the challenge of tackling complex quantum theories. Over decades, physics has evolved from classical mechanics to intricate frameworks like quantum field theory and quantum chromodynamics, creating both insight and paradox. This reflection emerges from a deep sense of curiosity, shared by many, to understand not just what the universe is made of but how these theoretical structures genuinely map onto reality.

2. The Crisis of Modern Physics: From Classical Mechanics to the Quark Hypothesis

Moving through physics from classical theories into high-energy particle models reveals a stark contrast: classical mechanics offers clarity and empiricism, while modern particle theories, such as quarks and gluons, often feel abstract and detached from observable reality. The shift to “smoking gun physics”—observing particle jets rather than the particles themselves—highlights a methodological divide. While high-energy collisions produce vivid images and data, we must question whether these indirect observations validate quarks, or merely add complexity to our models.

3. Historical Context: Quantum Numbers and the Evolution of the Standard Model

The 1960s and 70s were pivotal for particle physics, introducing quantum numbers like strangeness, charm, and beauty to account for unexplained phenomena in particle interactions. Figures like Murray Gell-Mann and Richard Feynman attempted to classify particles by assigning these numbers, essentially ad hoc solutions to match data with theoretical expectations. However, as experiments push the boundaries, new data shows that these quantum numbers often fail to predict actual outcomes consistently.

One of the key criticisms of this approach lies in the arbitrary nature of these quantum numbers. When certain decays were unobserved, strangeness was introduced as a “conservation law,” but when that proved insufficient, additional numbers like charm were added. The Standard Model has thus evolved not from fundamental truths, but as a patchwork of hypotheses that struggle to keep pace with experimental findings.

4. The Nobel Prize and the Politics of Scientific Recognition

Scientific recognition, especially through the Nobel Prize, has reinforced certain theories by celebrating theoretical advances sometimes over empirical confirmation. While groundbreaking work should indeed be recognized, the focus on theoretical predictions has, at times, overshadowed the importance of experimental accuracy and reproducibility. This dynamic may have inadvertently constrained the scope of mainstream physics, favoring elaborate but tenuous theories over simpler, empirically grounded explanations.

For example, Nobel Prizes have been awarded to proponents of the quark model and the Higgs boson long before we fully understand these particles’ empirical foundations. In doing so, the scientific community risks prematurely canonizing incomplete or even incorrect theories, making it challenging to revisit or overturn these assumptions without undermining established reputations.

5. Indirect Evidence: The Limits of Particle Accelerators

Particle accelerators, particularly at scales such as CERN’s Large Hadron Collider, have extended our observational reach, yet the evidence remains indirect. High-energy collisions create secondary particles and jets rather than isolated quarks or gluons. In a sense, we are not observing the fundamental particles but rather the “smoking gun” evidence they purportedly leave behind. The data produced are complex patterns and distributions, requiring interpretations laden with theoretical assumptions.

This approach raises a fundamental question: if a theory only survives through indirect evidence, can it be considered complete or even valid? High-energy experiments reveal that the more energy we input, the more complex the decay products become, yet we remain without direct evidence of quarks themselves. This “smoking gun” approach diverges from the empirical rigor demanded in classical physics and undermines the predictive power we might expect from a true theory of fundamental particles.

6. The Particle Zoo: A Growing Complexity

The “particle zoo” has expanded over decades, complicating rather than simplifying our understanding of matter. Initial hopes were that quantum numbers and conservation laws like strangeness would organize particles in a coherent framework, yet the resulting classification scheme has only grown more convoluted. Today, particles such as baryons, mesons, and leptons are grouped by properties derived not from first principles but from empirical fits to data, leading to ad hoc conservation laws that seem arbitrary.

The “strangeness” quantum number, for instance, was initially introduced to prevent certain reactions from occurring. Yet, rare reactions that violate this rule have been observed, suggesting that the rule itself is more of a guideline than a fundamental conservation law. This trend continued with the addition of quantum numbers like charm, beauty, and even bottomness, yet these additions have not resolved the core issue: our inability to explain why certain reactions occur while others do not.

7. Disequilibrium States: Beyond the Particle Concept

One possible perspective is to reclassify many “particles” not as fundamental entities but as disequilibrium states—transient structures that emerge from the interactions of more fundamental components. Viewing particles in this way offers a pathway back to a simpler, more intuitive model, where only stable particles like electrons, protons, and photons are foundational. Such a model could focus on electromagnetic fields and forces, with high-energy states representing temporary disequilibrium configurations rather than new particle species.

This perspective aligns well with the principle of statistical determinism. In the same way that classical oscillators eventually dampen and settle into stable states, high-energy disequilibrium states would be expected to decay, producing stable configurations over time. This model not only reduces the need for numerous quantum numbers but also sidesteps the requirement for exotic forces like the strong and weak nuclear forces, allowing the electromagnetic force to assume a central role.

8. Statistical Determinism and Quantum Reality

Heisenberg and Bohr’s interpretation of quantum mechanics suggests we should accept statistical determinism—systems governed by probabilistic rules where precise knowledge of individual events is inaccessible. This idea does not necessitate mystical randomness but acknowledges our limited ability to track initial conditions in high-energy environments. Probabilities emerge not from an intrinsic unpredictability but from our practical inability to fully specify a system’s state.

From this viewpoint, quarks and gluons, as well as the numerous quantum numbers assigned to unstable particles, are secondary descriptors rather than primary components of nature. Stable particles are the true constants, while all else is a function of high-energy interactions. This interpretation keeps quantum mechanics grounded in empirical reality and sidesteps the need for complex, unverifiable entities.

9. Conclusion: Toward a Pragmatic and Local Realist Approach

This reflection does not dismiss the importance of high-energy physics but advocates a return to fundamental principles. By focusing on empirical evidence, statistical determinism, and electromagnetic interactions, we can build a model that is both pragmatic and intuitive. We need not abandon quantum mechanics, but we should strive to ensure that its interpretations are consistent with the observable universe. Instead of introducing additional quantum numbers or forces, we should ask if these are placeholders for deeper, more coherent explanations yet to be discovered.

The journey of science is, at its core, a journey back to simplicity. If physics is to move forward, it may do so by revisiting foundational assumptions, clarifying what can be empirically tested, and developing a model of matter that resonates with the simplicity we find in classical theories. As research continues, it is this blend of skepticism, open-mindedness, and empirical rigor that will pave the way for meaningful discoveries.

On the quantum computing hype

1. The Wikipedia article on quantum computing describes a quantum computer as “a computer that exploits quantum -mechanical phenomena.” The rest of the article then tries to explain what these quantum-mechanical phenomena actually are.

Unfortunately, the article limits itself to the mainstream interpretation of these and, therefore, suffers from what I perceive to be logical and philosophical errors. Indeed, in the realistic interpretation of quantum mechanics that I have been developing, system wavefunctions are only useful to model our own uncertainty about the system. I subscribe to Hendrik Antoon Lorentz’s judgment at the last Solvay Conference under his leadership: there is no need whatsoever to elevate indeterminism to a philosophical principle. Not in science in general, and not in quantum mechanics in particular. I, therefore, think quantum mechanics cannot offer a substantially new computing paradigm.

Of course, one may argue that, for specific problems, some kind of three- or more-valued logic – rather than the binary or Boolean true/false dichotomy on which most logic circuits are based – may come in handy. However, such logic has already been worked out, and can be accessed using appropriate programming languages. Python and the powerful mathematical tools that come with it (Pandas, NumPy and SciPy) work great with ternary logic using a {true, false, unknown} or a {-1, 0, +1} set of logical values rather than the standard {0, 1} Boolean set. The Wikipedia article on three-valued logic is worth a read and, despite the rather arcane nature of the topic, much better written than the mentioned article: have a look at how operators are used on these three-valued sets in meaningful algebras or logical models, such as that of Kleene, Priest or Lukasiewicz.

2. One may, of course, argue that, even when there is probably no such thing as a new logical quantum computing model or logic, quantum technology may offer distinct advantages when it comes to storage of data about this or that state or, one day, lead to devices with faster clock and/or bus speeds. That appears to be a pipedream too:

  • To keep, say, an electron in this or that spin state, one must create and steady an electromagnetic field – usually one does so in a superconducting environment, which makes actual mechanical devices used for quantum computing (qubits) look like the modern-day equivalent of Babbage’s analytical machine. In my not-so-humble view, such devices will never ever achieve the sheer material performance offered by current nanometer-scale MOSFETs.  

  • As for bus or transmission speeds, quantum theory does not come with a new theory of charge propagation and, most importantly, is fundamentally flawed in its analysis of how signals actually propagate in, say, a lattice structure. I refer to one of my papers here (on electron propagation in a lattice), in which I deconstruct Feynman’s analysis of the concept of the free and effective mass of an electron. Hence, for long-distance transmission of signals, optical fiber cannot be beaten. For short-distance transmission of signals (say, within an electrical circuit, I refer to the above-mentioned nano-technology which continues to revolutionize the chip industry.

Brussels, 4 July 2023