We climbed a mountain—step by step—post by post. :-) We have reached the top now, and the view is gorgeous. We understand Schrödinger’s equation, which describes how amplitudes propagate through space-time. It’s the quintessential quantum-mechanical expression. Let’s enjoy now, and deepen our understanding by introducing the concept of (quantum-mechanical) *operators*. Let me do so by first making a remark on notation. You’ll remember we wrote Schrödinger’s famous equation as:

However, you may have seen the following inscription on his bust, or on his grave, or—somewhat less morbid or deferential—just the following formula, which summarizes the whole expression as:

The H in this expression is, of course, *not *the Hamiltonian *matrix*, but an **operator**. So the same symbol (H) is used to denote two different things. To distinguish the two, we should use the *hat *symbol, so as to distinguish the matrix (e.g. A) from the operator (Â). However, it’s just like studying statistics: the *hat *symbol is supposed to distinguish some *estimator function* (*â*) from the parameter itself (α), or from the *estimate* of the parameter, i.e. the *observation* (*a*). However, you’ll surely remember the *hat *disappeared pretty quickly in your statistics course, because the *context *is usually enough to see what’s being meant. So… Well… I’ll be sloppy as well here, if only because the *WordPress* editor only offers very few symbols with a hat! :-)

Now, you’ll note the H operator in the expression above is pretty monstrous as it’s, obviously, identical to:

As you can see, this H operator actually consists of *two *other operators: (1) the ∇^{2 }operator, which you know (∇^{2 }= ∂/∂x^{2 }+ ∂/∂y^{2 }+ ∂/∂z^{2}), and (2) the V(*x*, *y*, *z*) ‘operator’, which—in this particular context—just means: “multiply with V”. [Needless to say, V is the *potential* here, and so it captures the presence of external force fields.]

So… Well… ** This** H here surely

*looks*

**very**different than the quantum-mechanical Hamiltonian operator we discussed when dealing with a

*finite*set of base states:

*H was nothing but the Hamiltonian*

**that***matrix*operating on some state indeed.

Having said that, it shouldn’t surprise you if I say that, despite the fact that they *look *so different, these two operators are actually *equivalent*: the only difference is that one is designed to operate on a (state) *vector*, while the other is designed to operate on a continuous *function*. Their interpretation is similar, as evidenced from the fact that both are being referred to as the *energy operator *in quantum physics.

So… Yes… Let’s talk about that. But let’s first review the basics.

**What is all that ***psi*-chology?

*psi*-chology?

We’ll need to go from what is referred to as *matrix mechanics* to what is referred to as *wave mechanics*. So… Well… Let’s start with matrix mechanics. The matrix-mechanical approach is summarized in that set of Hamiltonian equations which, by now, you know so well:

You understand this equation, but, even then, it’s always good to remind oneself of the description of a ** state**:

|ψ〉 is the *state *of a system, like the *ground state *of a hydrogen atom, or one of its many *excited *states. You also know that the *lifetime* of a system in an excited state is usually short: some *spontaneous* or *induced* emission of a quantum of energy (i.e. a *photon*) will ensure that the system quickly returns to a less excited state, or to the ground state itself. However, that doesn’t impact the analysis here: we’re looking at the state of the system *at some point in time* here. That’s all. In fact, that’s why we introduce the concept of operators: the state of the system will, inevitably, change—as time goes by.

That’s clear enough. However, I should warn you here. There’s this potential confusion. It’s caused by the ubiquity of the ψ symbol (i.e. the Greek letter * psi*). It’s really something

*psi*-chological. :-) In

*matrix mechanics*, our ψ would just denotes a

*state*of a

*system*, which could be an atom, which we’d describe by the orbital(s) of the electron(s) around it. In this regard, I found the following illustration from Wikipedia particularly helpful: the green orbitals show

*excitations*of

*copper*(Cu) orbitals on a CuO

_{2 }plane. [The two big arrows just illustrate the principle of X-ray spectroscopy, so it’s an X-ray

*probing*the structure of the material.]

So… Well… We’d write ψ as |ψ〉 just to remind ourselves we’re talking of some *state *of the** system** indeed. However, quantum physicists will also use the

*psi*symbol to denote some specific

*C*

_{i}*amplitude*(or

*coefficient*) in that |ψ〉 = ∑|

*i*〉

*C*formula above. To be specific, they’d replace the base states |

_{i}*i*〉 by the continuous position variable

*x*, and they would write the following:

*C _{i}* = ψ(

*i*=

*x*) = ψ(

*x*) =

*C*

_{ψ}(

*x*) =

*C*(

*x*) = 〈

*x*|ψ〉

In fact, that’s just like writing:

φ(*p*) = 〈 mom *p* | ψ 〉 = 〈*p*|ψ〉 = *C*_{φ}(*p*) = *C*(*p*)

What they’re doing here, is (1) reduce the ‘*system*‘ to a ‘*particle*‘ once more (which is OK, as long as you know what you’re doing) and (2) they basically state the following:

If a particle is in some state |ψ〉, then we can associate some *wavefunction *ψ(*x*) or φ(*p*)—with it, and that wavefunction will represent the *amplitude *for the system (i.e. our particle) to be at *x*, or to have a momentum that’s equal to *p.*

So they should have written χ(x) instead of ψ(*x*), I feel, so as to avoid confusion: one should *not* use the same symbol for the |ψ〉 state and the ψ(*x*) wavefunction. The point is: the *position *or the *momentum*, or even the *energy*, are * properties* of the |ψ〉 state and, therefore, it’s really confusing to use the same symbol

*psi*(ψ) to describe (1) the

*state*, and (2) the

*wavefunction*of just

*one*of the various properties of that state (in this case: its

*position*). In fact, that’s what this post is all about: it’s about how to describe certain

*properties*of the system. Of course, we’re talking quantum mechanics here and, hence,

*uncertainty*, and, therefore, we’re going to talk about the

*average*position, energy, momentum, etcetera that’s associated with a particular

*state*of a system, or—as we’ll keep things

*very*simple—the properties of a ‘particle’. Think of an electron in some orbital, indeed! :-)

So let’s now look at that set of Hamiltonian equations once again:

Looking at it *carefully* – so just look at it once again! :-) – and thinking about what we did when going from the discrete to the continuous setting, we can now understand we should write the following for the continuous case:

Of course, combining Schrödinger’s equation with the expression above implies the following:

Now how can we relate that integral to the expression on the right-hand side? I’ll have to disappoint you here, as it requires a lot of math to transform that integral. It requires writing H(*x*, *x’*) in terms of rather complicated functions, including – you guessed it, didn’t you? – Dirac’s delta function. Hence, I assume you’ll believe me if I say that the matrix- and wave-mechanical approaches *are* actually equivalent. In any case, if you’d want to check it, you can always read Feynman yourself. :-)

Now, I wrote this post to talk about quantum-mechanical *operators*, so let me do that now.

**Quantum-mechanical operators**

You know the concept of an operator. As mentioned above, we should put a little *hat *(^) on top of our Hamiltonian operator, so as to distinguish it from the matrix itself. However, as mentioned above, the difference is usually quite clear from the context. Our operators were all matrices so far, and we’d write the matrix elements of, say, some operator A, as:

A_{ij} ≡ 〈 *i* | A | *j* 〉

The whole matrix itself, however, would usually not act on a base state but… Well… Just on some state ψ, to produce some new state φ, and so we’d write:

| φ 〉 = A | ψ 〉

Of course, we’d have to *describe *| φ 〉 in terms of the (same) set of base states and, therefore, we’d expand this expression into something like this:

You get the idea. I should just add one more thing. You know this important property of amplitudes: the 〈 ψ | φ 〉 amplitude is the *complex conjugate *of the 〈 φ | ψ 〉 amplitude. It’s got to do with time reversibility, because the complex conjugate of *e*^{−iθ }= *e*^{−i(ω·t−k·x) }is equal to *e*^{iθ }= *e*^{i(ω·t−k·x)}, so we’re just reversing the *x*– and *t-*direction. We write:

〈 ψ | φ 〉 = 〈 φ | ψ 〉*

Now what happens if we want to take the complex conjugate when we insert a matrix, so when writing 〈 φ | A | ψ 〉 instead of 〈 φ | ψ 〉, this rules becomes:

〈 φ | A | ψ 〉* = 〈 ψ | A† | φ 〉

The *dagger *symbol denotes the *conjugate transpose*, so A† is an operator whose matrix elements are equal to A_{ij}† = A_{ji}*. Now, it may or may not happen that the A† matrix is actually equal to the original A matrix. In that case – and *only *in that case – we can write:

〈 ψ | A | φ 〉 = 〈 φ | A | ψ 〉*

We then say that A is a ‘self-adjoint’ or ‘Hermitian’ operator. That’s just a definition of a property, but many quantum-mechanical operators are actually Hermitian. In any case, we’re well armed now to discuss some *actual *operators, and we’ll start with that *energy *operator.

**The energy operator (H)**

We know the state of a *system *is described in terms of a set of *base* states. Now, our analysis of *N*-state systems showed we can always describe it in terms of a *special *set of base states, which are referred to as the **states of definite energy **because… Well… Because they’re associated with some *definite *energy. In that post, we referred to these energy levels as *E*_{n }(** n** = I, II,…

**N**). We used boldface for the subscript n (so we wrote

**n**instead of n) because of these

*Roman*numerals. With each energy level, we could associate a base state, of

*definite energy*indeed, that we wrote as |

**n**〉. To make a long story short, we summarized our results as follows:

- The energies
*E*_{I},*E*_{II},…,*E*,…,_{n}*E*_{N }are theof the Hamiltonian matrix H.**eigenvalues** - The state vectors |
**n**〉 that are associated with each energy*E*, i.e. the set of vectors |_{n}**n**〉, are the corresponding.*eigenstates*

We’ll be working with some more subscripts in what follows, and these Roman numerals and the boldface notation are somewhat confusing (if only because I don’t want you to think of these subscripts as *vectors*), we’ll just denote *E*_{I}, *E*_{II},…, *E*** _{n}**,…,

*E*

_{N }as

*E*

_{1},

*E*

_{2},…,

*E*,…,

_{i}*E*

_{N}, and we’ll

*number*the states of definite energy accordingly, also using some Greek letter so as to clearly distinguish them from all our

*Latin*letter symbols: we’ll write these states as: |η

_{1}〉, |η

_{1}〉,… |η

_{N}〉. [If I say, ‘we’, I mean Feynman of course. You may wonder why he doesn’t write |E

*〉, or |ε*

_{i}*〉. The answer is: writing |E*

_{i}_{n}〉 would cause confusion, because this state will appear in expressions like: |E

*〉E*

_{i}*, so that’s the ‘product’ of a state and the associated*

_{i}*scalar*. Too confusing. As for using η (

*eta*) instead of ε (

*epsilon*) to denote something that’s got to do with

**nergy… Well… I guess he wanted to keep the**

*e**resemblance*with the

**n**, and then the Ancient Greek apparently did use this η letter for a sound like ‘

*e*‘ so… Well… Why not? Let’s get back to the lesson.]

Using these base states of definite energy, we can write the state of the system as:

|ψ〉 = ∑ |η* _{i}*〉 C

_{i }= ∑ |η

*〉〈η*

_{i}

_{i}*|*ψ〉

*over all*

_{ }*i*(

*i*= 1, 2,… , N)

Now, we didn’t talk all that much about what these base states actually *mean *in terms of measuring something but you’ll believe if I say that, when *measuring *the energy of the system, we’ll always measure one *or* the other *E*_{1}, *E*_{2},…, *E _{i}*,…,

*E*

_{N}value. We’ll never measure something in-between: it’s either-or. Now, as you know, measuring something in quantum physics is supposed to be destructive but… Well… Let us

*imagine*we could make a thousand measurements to try to determine the

*average*energy of the system. We’d do so by counting the number of times we measure

*E*

_{1}(and of course we’d denote that number as N

_{1}),

*E*

_{2},

*E*

_{3}, etcetera. You’ll agree that we’d measure the average energy as:

However, measurement is destructive, and we actually *know *what the *expected value *of this ‘average’ energy will be, because we know the *probabilities* of finding the system in a particular base state. That probability is equal to the *absolute *square of that C* _{i }*coefficient above, so we can use the P

*= |*

_{i }*C*|

_{i}^{2}formula to write:

〈*E _{av}*〉 = ∑ P

_{i }

*E*over all

_{i}*i*(

*i*= 1, 2,… , N)

Note that this is a rather general formula. It’s got nothing to do with quantum mechanics: if A* _{i}* represents the

*possible*values of some quantity A, and P

*is the probability of getting that value, then (the expected value of) the average A will also be equal to 〈*

_{i}*A*〉 = ∑ P

_{av}*A*

_{i }*. No rocket science here! :-) But let’s now apply our quantum-mechanical formulas to that 〈*

_{i}*E*〉 = ∑ P

_{av}_{i }E

*formula. [Oh—and I apologize for using the same angle brackets 〈 and 〉 to denote an expected value here—sorry for that! But it’s what Feynman does—and other physicists! You see: they don’t*

_{i}*really*want you to understand stuff, and so they often use very confusing symbols.] Remembering that the absolute square of a complex number equals the product of that number and its complex conjugate, we can re-write the 〈

*E*〉 = ∑ P

_{av}_{i }E

*formula as:*

_{i}〈*E _{av}*〉 = ∑ P

_{i }

*E*= ∑ |

_{i}*C*|

_{i}^{2 }

*E*= ∑

_{i}*C**

_{i}*C*= ∑

_{i }E_{i}*C**

_{i }*C*= ∑ 〈ψ|η

_{i }E_{i}_{i}〉〈η

_{i}*|*ψ〉

*E*= ∑ 〈ψ|η

_{i }_{i}〉

*E*

_{i}〈η

_{i}*|*ψ〉 over all

*i*

Now, you know that Dirac’s *bra-ket *notation allows numerous manipulations. For example, what we could do is take out that ‘common factor’ 〈ψ|, and so we may re-write that monster above as:

〈*E _{av}*〉 = 〈ψ| ∑ η

_{i}〉

*E*

_{i}〈η

_{i}*|*ψ〉 = 〈ψ|φ〉, with |φ〉 = ∑ |η

_{i}〉

*E*

_{i}〈η

_{i}*|*ψ〉 over all

*i*

* Huh? *Yes. Note the difference between |ψ〉 = ∑ |η

*〉 C*

_{i}_{i }= ∑ |η

*〉〈η*

_{i}

_{i}*|*ψ〉 and |φ〉 = ∑ |η

_{i}〉

*E*

_{i}〈η

_{i}*|*ψ〉. As Feynman puts it: φ is just some ‘

*cooked-up*‘ state which you get by taking each of the base states |η

_{i}〉 in the amount

*E*

_{i}〈η

_{i}*|*ψ〉 (as opposed to the 〈η

_{i}*|*ψ〉 amounts we took for ψ).

I know: you’re getting tired and you wonder why we need all this stuff. Just hang in there. We’re almost done. I just need to do a few more unpleasant things, one of which is to remind you that this business of the energy states being *eigenstates* (and the energy levels being *eigenvalues*) of our Hamiltonian matrix (see my post on *N*-state systems) comes with a number of interesting properties, including this one:

H |η_{i}〉 = *E*_{i}|η_{i}〉 = |η_{i}〉*E*_{i}

Just think about it: on the left-hand side, we’re multiplying a matrix with a (base) state vector, and on the left-hand side we’re multiplying it with a *scalar*. So our |φ〉 = ∑ |η_{i}〉*E*_{i}〈η_{i}*|*ψ〉 sum now becomes:

|φ〉 = ∑ H |η_{i}〉〈η_{i}*|*ψ〉 over all *i *(*i* = 1, 2,… , N)

Now we can manipulate that expression some more so as to get the following:

|φ〉 = H ∑|η_{i}〉〈η_{i}*|*ψ〉 = H*|*ψ〉

Finally, we can re-combine this now with the 〈*E _{av}*〉 = 〈ψ|φ〉 equation above, and so we get the fantastic result we wanted:

〈*E _{av}*〉 = 〈 ψ | φ 〉 = 〈 ψ | H

*|*ψ 〉

* Huh? *Yes!

**To get the average energy, you operate on**|

**ψ**〉

**with H, and then you multiply the result with**〈

**ψ**|. It’s a beautiful formula. As Feynman notes, the new formula for the average energy is not only pretty but also useful, because now we don’t need to say anything about any particular set of base states. We don’t even have to know all of the possible energy levels. When we go to calculate, we’ll need to describe our state in terms of

*some*set of base states, but if we know the Hamiltonian matrix for

*that*set, we can get the average energy.

Of course, you know that, if you’ve the Hamiltonian, you always have everything, so… Well… Yes. You’re right: it’s less of a big deal than it seems. Having said that, the whole development above is very interesting because of something else: we can easily *generalize *it for other physical measurements. I call it the ‘average value’ operator idea, but you won’t find that term in any textbook. :-) Let me explain the idea.

**The average value operator (A)**

The development above illustrates how we can relate a *physical observable*, like the (average) energy (E), to a quantum-mechanical operator (H). Now, the development above can easily be generalized to *any observable that would be proportional to the energy*. It’s perfectly reasonable, for example, to assume the angular momentum – as measured in some direction, of course, which we usually refer to as the z-direction – would be proportional to the energy, and so then it would be easy to define a new operator L_{z}, which we’d define as the operator of the z-component of the angular momentum ** L**. [I know… That’s a bit of a long name but… Well… You get the idea.] So we can write:

〈*L _{z}*〉

*= 〈 ψ | L*

_{av}_{z }

*|*ψ 〉

In fact, further generalization yields the following grand result:

If a physical observable A is related to a suitable quantum-mechanical operator Â, then the average value of A for the state | ψ 〉 is given by:

〈*A*〉* _{av}* = 〈 ψ | Â

_{ }

*|*ψ 〉 = 〈 ψ | φ 〉 with | φ 〉 = Â

_{ }

*|*ψ 〉

At this point, you may have second thoughts, and wonder: ** what** state | ψ 〉? The answer is: it doesn’t matter. It can be any state,

*as long as we’re able to describe in terms of a chosen set of base states*. :-)

OK. So far, so good. The next step is to look at how this works for the continuity case.

**The energy operator for wavefunctions (H)**

We can start thinking about the *continuous *equivalent of the 〈*E _{av}*〉 = 〈ψ|H

*|*ψ〉 expression by first expanding it. We write:

You know the continuous equivalent of a sum like this is an integral, i.e. an *infinite *sum. Now, because we’ve got *two *subscripts here (*i* and *j*), we get the following *double *integral:

Now, I did take my time to walk you through Feynman’s derivation of the energy operator for the *discrete *case, i.e. the operator when we’re dealing with *matrix mechanics*, but I think I can simplify my life here by just copying Feynman’s succinct development:

Done! Given a wavefunction ψ(x), we get the average energy by doing that integral above. Now, the quantity in the braces of that integral can be written as that operator we introduced when we started this post:

So now we can write that integral much more elegantly. It becomes:

〈*E*〉* _{av}* = ∫ ψ*(

*x*)

*H*ψ(

*x*) d

*x*

You’ll say that doesn’t look like 〈*E _{av}*〉 = 〈 ψ | H

*|*ψ 〉! It does. Remember that 〈 ψ | =

*|*ψ 〉*. :-)

*Done!*I should add one qualifier though: the formula above assumes our wavefunction has been normalized, so all probabilities add up to one. But that’s a minor thing. The only thing left to do now is to generalize to three dimensions. That’s easy enough. Our expression becomes a *volume *integral:

〈*E*〉* _{av}* = ∫ ψ*(

**)**

*r**H*ψ(

**) d**

*r**V*

Of course, d*V* stands for d*Volume *here, not for any potential energy, and, of course, once again we assume all probabilities over the volume add up to 1, so all is normalized. ** Done! ** :-)

We’re almost done with this post. What’s left is the *position *and *momentum *operator. You may think this is going to another lengthy development but… Well… It turns out the analysis is remarkably simple. Just stay with me a few more minutes and you’ll have earned your degree. :-)

**The position operator (***x*)

*x*)

The thing we need to solve here is really easy. Look at the illustration below as representing the probability density of some particle being at *x*. Think about it: what’s the average position?

Well? What? The (expected value of the) average position is just this simple integral: 〈*x*〉* _{av}* = ∫

*x*P(

*x*) d

*x*, over all the whole range of possible values for x. :-) That’s all. Of course, because P(

*x*) = |ψ(

*x*)|

^{2}=ψ*(

*x*)·ψ(

*x*), this integral now becomes:

〈*x*〉* _{av}* = ∫ ψ*(

*x*)

*x*ψ(

*x*) d

*x*

That looks *exactly *the same as 〈*E*〉* _{av}* = ∫ ψ*(

*x*)

*H*ψ(

*x*) d

*x*, and so we can look at

*x*as an operator too!

** Huh?** Yes. It’s an extremely simple operator: it

*just means “multiply by*

*x*“. :-)

I know you’re shaking your head now: is it *that *easy? It is. Moreover, the ‘matrix-mechanical equivalent’ is equally simple but, as it’s getting late here, I’ll refer you to Feynman for that. :-)

**The momentum operator (***p*_{x}**)**

Now we want to calculate the average momentum of, say, some electron. What integral would you use for that? […] Well… *What?* […] It’s easy: it’s the same thing as for *x*. We can just substitute replace *x *for *p *in that 〈*x*〉* _{av}* = ∫

*x*P(

*x*) d

*x*formula, so we get:

〈*p*〉* _{av}* = ∫

*p*P(

*p*) d

*p*, over all the whole range of possible values for

*p*

Now, you might think the rest is equally simple, and… Well… It actually *is *simple but there’s one additional thing in regard to the need to normalize stuff here. You’ll remember we defined a *momentum *wavefunction (see my post on the Uncertainty Principle), which we wrote as:

φ(p) = 〈 mom *p* | ψ 〉

Now, in the mentioned post, we related this *momentum* wavefunction to the particle’s ψ(x) = 〈x|ψ〉 wavefunction—which we should actually refer to as the *position *wavefunction, but everyone just calls it *the *particle’s wavefunction, which is a bit of a misnomer, as you can see now: a wavefunction describes some *property *of the system, and so we can associate *several *wavefunctions with the same system, really! In any case, we noted the following there:

- The two probability density functions, φ(p) and ψ(
*x*), look pretty much the same, but the*half-width*(or standard deviation) of one was inversely proportional to the half-width of the other. To be precise, we found that the constant of proportionality was equal to ħ/2, and wrote that relation as follows: σ_{p}_{ }= (ħ/2)/σ_{x}. - We also found that, when using a regular normal distribution function for ψ(
*x*), we’d have to normalize the probability density function_{x}^{2})^{−1/2 }in front of the exponential.

Now, it’s a bit of a complicated argument, but the upshot is that we cannot just write what we usually write, i.e. P* _{i}* = |

*C*|

_{i}^{2 }or P(

*x*) = |ψ(

*x*)|

^{2}. No. We need to put a normalization factor in front, which combines the two factors I mentioned above. To be precise, we have to write:

P(*p*) = |〈*p*|ψ〉|^{2}/(2πħ)

So… Well… Our 〈*p*〉* _{av}* = ∫

*p*P(

*p*) d

*p*integral can now be written as:

〈*p*〉* _{av}* = ∫ 〈ψ|

*p*〉

*p*〈

*p*|ψ〉 d

*p*/(2πħ)

So that integral is totally like what we found for 〈*x*〉* _{av}* and so… We could just leave it at that, and say we’ve solved the problem. In that sense, it

*is*easy. However, having said that, it’s obvious we’d want some solution that’s written in terms of ψ(

*x*), rather than in terms of φ(

*p*), and that requires some more manipulation. I’ll refer you, once more, to Feynman for that, and I’ll just give you the result:

So… Well… I turns out that the momentum operator – which I tentatively denoted as *p*_{x} above – is *not *so simple as our position operator (*x*). Still… It’s not *hugely *complicated either, as we can write it as:

*p*_{x} ≡ (ħ/*i*)·(∂/∂*x*)

Of course, the *purists *amongst you will, once again, say that I should be more careful and put a *hat *wherever I’d need to put one so… Well… You’re right. I’ll wrap this all up by copying Feynman’s overview of the operators we just explained, and so he *does *use the fancy symbols. :-)

Well, folks—that’s it! Off we go! You know all about quantum physics now! We just need to work ourselves through the exercises that come with Feynman’s *Lectures*, and then you’re ready to go and bag a degree in physics somewhere. So… Yes… That’s what *I* want to do now, so I’ll be silent for quite a while now. Have fun! :-)