**Pre-script** (dated 26 June 2020): I have come to the conclusion one does not need all this hocus-pocus to explain masers or lasers (and two-state systems in general): classical physics will do. So no use to read this. Read my papers instead. ðŸ™‚

**Original post**:

As I skipped the mathematical argumentsÂ in my previous post so as to focus on the essential *results* only, I thought it would be good to complement that post*Â *byÂ looking at the math once again, so as to ensure we understand what it is that we’re doing. So let’s do that now. We start with the easy situation: free space.

**The two-state system in free space**

We started with an ammonia molecule in free space, i.e. we assumed there were no external force fields, like a gravitational or an electromagnetic force field. Hence, the picture was as simple as the one below: the nitrogen atom could be ‘up’ or ‘down’ with regard to its spin around its axis of symmetry.

It’s important to note that this ‘up’ or ‘down’ direction is defined in regard to the molecule itself, i.e.Â *notÂ *in regard to someÂ *external*Â reference frame. In other words, the reference frame is that of the molecule itself. For example, if I flip the illustration above â€“ like below â€“ then we’re still talking the same states, i.e. the molecule is still in state 1 in the image on theÂ left-handÂ sideÂ and it’s still in state 2 in the image on theÂ right-hand side*.**Â *

We then modeled the *uncertaintyÂ *about its state by associatingÂ *twoÂ *different energy levels with the molecule:Â E_{0 }+ A and E_{0Â }âˆ’ A. The idea is that the nitrogen atom needs to tunnel through a potential barrier to get to the other side of the plane of the hydrogens, and that requires energy. At the same time, we’ll show the two energy levels are effectively associated with an ‘up’ or ‘down’ direction of the electric dipole moment of the molecule. So that resembles the two spin states of an electron, which we associated with the +Ä§/2 and âˆ’Ä§/2 energies respectively. So ifÂ E_{0 }would be zero (we can always take another reference point, remember?), then we’ve got the same thing: two energy levels that are separated by someÂ *definiteÂ *amount: that amount is 2A for the ammonia molecule, and Ä§ when we’re talking quantum-mechanical spin. I should make a last note here, before I move on: note that these energies only make sense in the presence of some external field, because the + and âˆ’ signs in the E_{0 }+ A and E_{0Â }âˆ’ A and +Ä§/2 and âˆ’Ä§/2 expressions make sense only *with regard to some external direction*Â defining what’s ‘up’ and what’s ‘down’ really. But I am getting ahead of myself here. Let’s go back to free space: no external fields, so what’s ‘up’ or ‘down’ is completely random here. ðŸ™‚

Now, we also know an energy level can beÂ associated with a complex-valued wavefunction, or an *amplitude* as we call it. To be precise, we can associate it with the genericÂ *aÂ·e*^{âˆ’(i/Ä§)Â·(EÂ·t âˆ’ pâˆ™x)}^{Â }expression which you know so well by now. Of course, Â as the reference frame is that of the molecule itself, its momentum is zero, so the **p**âˆ™**x**Â term in theÂ *aÂ·e*^{âˆ’(i/Ä§)Â·(EÂ·t âˆ’ pâˆ™x)}^{Â }expression vanishes and the wavefunction reduces to *aÂ·e*^{âˆ’iÂ·Ï‰Â·tÂ }=Â *aÂ·e*^{âˆ’(i/Ä§)Â·EÂ·t}, with Ï‰ = E/Ä§. In other words, the energy level determines the *temporalÂ *frequency, or the temporal variationÂ (as opposed to the spatial frequency or variation), of the amplitude.

We then had to find the *amplitudes*Â C_{1}(t) = âŒ© 1 | Ïˆ âŒª and C_{2}(t) =âŒ© 2 | Ïˆ âŒª, so that’s the amplitude to be in state 1 *orÂ *state 2 respectively. In my post on the Hamiltonian, I explained why the *dynamicsÂ *of a situation like this can be represented by the following set of differential equations:

As mentioned, the C_{1Â }and C_{2 }functions evolve in time, and so we should write them as C_{1Â }= C_{1}(t) and C_{2Â }= C_{2}(t) respectively. In fact, our Hamiltonian coefficients may also evolve in time, which is why it may be very difficult to solve those differential equations! However, as I’ll show below, one usually assumes they are constant, and then one makes informed guesses about them so as to find a solution that makes sense.

Now, I should remind you here of something you surely know: if C_{1Â }and C_{2Â }are solutions to this set of differential equations, then the superposition principle tells us that any linear combination*Â *aÂ·C_{1 }+ bÂ·C_{2Â }will also be a solution. So we need one or more extra conditions, usually someÂ *startingÂ *condition, which we can combine with a *normalizationÂ *condition,Â so we can get someÂ *uniqueÂ *solution that makes sense.

The H_{ijÂ }coefficients are referred to asÂ *HamiltonianÂ *coefficients and, as shown in the mentioned post, the H_{11Â }and H_{22Â }coefficients are related to theÂ amplitude*Â *of the molecule *staying* in state 1 and state 2 respectively, while the H_{12Â }and H_{21Â }coefficients are related to the amplitude of the molecule going from state 1 to state 2 and vice versa. Because of the perfect symmetry of the situation here, it’s easy to see that H_{11 }should equal H_{22Â }, and that H_{12Â }and H_{21Â }should also be equal to each other. Indeed, Nature doesn’t care what we call state 1 or 2 here: as mentioned above, we did *notÂ *defineÂ the ‘up’ and ‘down’ direction with respect to some external direction in space, so the molecule can have any orientation and, hence, switching the *i* an *j* indices should not make any difference.Â So that’s one clue, at least, that we can use to solve those equations: the perfect symmetry of the situation and, hence, the perfect symmetry of the Hamiltonian coefficientsâ€”in *thisÂ *case, at least!

The other clue is to think about the solution if we’d *not* haveÂ *two *states*Â *butÂ *one state only*. In that case, we’d need to solve *i*Ä§Â·[dC_{1}(t)/dt] = H_{11}Â·C_{1}(t). That’s simple enough, because you’ll remember that the exponential function is its own derivative. To be precise, we write: d(*a*Â·*e*^{iÏ‰t})/dt = *a*Â·d(*e*^{iÏ‰t})/dt = *a*Â·*i*Ï‰Â·*e*^{iÏ‰t}, and please note that *aÂ *can be anyÂ *complexÂ *number: we’re not necessarily talking a real number here! In fact, we’re likely to talk complex coefficients, and we multiply with some other complex number (*i*Ï‰) anyway here!Â So if we write *i*Ä§Â·[dC_{1}/dt] = H_{11}Â·C_{1}Â asÂ dC_{1}/dt = âˆ’(*i*/Ä§)Â·H_{11}Â·C_{1}Â (remember:Â *i*^{âˆ’1}Â = 1/*i*Â = âˆ’*i*), then it’s easy to see that the C_{1Â }=Â *a*Â·e^{â€“(i/Ä§)Â·H11Â·tÂ }functionÂ is the general solution for this differential equation. Let me write it out for you, just to make sure:

dC_{1}/dt =Â d[*a*Â·*e*^{â€“(i/Ä§)H11t}]/dt = *a*Â·d[*e*^{â€“(i/Ä§)H11t}]/dt = â€“*a*Â·(*i*/Ä§)Â·H_{11}Â·*e*^{â€“(i/Ä§)H11t}

= â€“(*i*/Ä§)Â·H_{11}Â·*a*Â·*e*^{â€“(i/Ä§)H11tÂ }=Â âˆ’(*i*/Ä§)Â·H_{11}Â·C_{1}

Of course, that reminds us of our generic wavefunction*Â aÂ·e*^{âˆ’(i/Ä§)Â·E0Â·t}Â wavefunction: we only need to equate H_{11}Â with E_{0}Â and we’re done! Hence, in a one-state system, the Hamiltonian coefficient is, quite simply, equal to theÂ *energyÂ *of the system. In fact, that’s a result can be generalized, as we’ll see below, and so that’s why Feynman says the Hamiltonian ought to be called theÂ ** energy matrix**.

In fact, we actually may have two states that are entirely uncoupled, i.e. a system in which there is no dependence ofÂ C_{1}Â onÂ C_{2Â }and vice versa. In that case, the two equations reduce to:

*i*Ä§Â·[dC_{1}/dt] = H_{11}Â·C_{1}Â and* i*Ä§Â·[dC_{2}/dt] = H_{22}Â·C_{2}

These do *notÂ *form a coupled system and, hence, theirÂ solutions are independent:

C_{1}(t)* =* *a*Â·e^{â€“(i/Ä§)Â·H11Â·tÂ }and C_{2}(t)_{Â }= *b*Â·e^{â€“(i/Ä§)Â·H22Â·tÂ }

The symmetry of the situation suggests we should equate *a* and *b*, and then the normalization condition says that the probabilities have to add up to one, so |C_{1}(t)|^{2Â }+ |C_{2}(t)|^{2Â }= 1, so we’ll find that *aÂ *=Â *bÂ *= 1/âˆš2.

OK. That’s simple enough, and this story has become quite long, so we should wrap it up. The two ‘clues’ â€“ about symmetry and about the Hamiltonian coefficients being energy levels â€“ lead Feynman to suggest that the Hamiltonian matrix for *this*Â particular case*Â *should be equal to:

Why? Well… It’s just one of Feynman’s clever guesses, and it yields probability functions that makes sense, i.e. they actually describe something *real*. That’s all. ðŸ™‚ I am only half-joking, because it’s a trial-and-error process indeed and, as I’ll explain in a separate section in this post, one needs to be aware of the various approximations involved when doing this stuff. So let’s be explicit about the reasoning here:

- We know that H
_{11 }= H_{22Â }= E_{0Â }if the two states would be identical. In other words, if we’d haveÂ*only one*state, rather than two â€“ i.e. if H_{12Â }and H_{21Â }would be zero â€“ then we’d just plug that in. So that’s what Feynman does. So that’s whatÂ*weÂ*do here too! ðŸ™‚ - However, H
_{12Â }and H_{21Â }are*notÂ*zero, of course, and so assume there’s some*amplitudeÂ*to go from one position to the other by tunneling through the energy barrier and flipping to the other side. Now, we need to assignÂ*someÂ*value to that amplitude and so we’ll just assumeÂ thatÂ the energy that’s needed for the nitrogen atom to tunnel through the energy barrier and flip to the other side is equal to A. So we equate H_{12Â }and H_{21}Â withÂ âˆ’A.

Of course, you’ll wonder: whyÂ *minus *A? Why wouldn’t we tryÂ H_{12Â }= H_{21}Â = A? Well… I could say that a particle usually *loses* potential energy as it moves from one place to another, but… Well… Think about it. Once it’s through, it’s through, isn’t it? And so then the energy is just E_{0Â }again. Indeed, if there’s no external field, the + or âˆ’ sign is quite arbitrary. So what do we choose? The answer is: when considering our molecule in free space, it doesn’t matter. Using +A or âˆ’A yields the same probabilities. Indeed, let me give you the *amplitudes*Â we get for H_{11 }= H_{22Â }= E_{0Â }and H_{12Â }and H_{21}Â = âˆ’A:

- C
_{1}(t) = âŒ© 1 | Ïˆ âŒª = (1/2)Â·*e*^{âˆ’(i/Ä§)Â·(E0Â âˆ’ A)Â·tÂ }+ (1/2)Â·*e*^{âˆ’(i/Ä§)Â·(E0Â + A)Â·tÂ }=*e*^{âˆ’(i/Ä§)Â·E0Â·t}Â·cos[(A/Ä§)Â·t] - C
_{2}(t) = âŒ© 2 | Ïˆ âŒª = (1/2)Â·*e*^{âˆ’(i/Ä§)Â·(E0Â âˆ’ A)Â·tÂ }â€“ (1/2)Â·*e*^{âˆ’(i/Ä§)Â·(E0Â + A)Â·tÂ }=*i*Â·*e*^{âˆ’(i/Ä§)Â·E0Â·t}Â·sin[(A/Ä§)Â·t]

[In case you wonder how we go from those exponentials to a simple sine and cosine factor, remember that the sum of complex conjugates, i.eÂ *e*^{iÎ¸Â }+Â *e*^{âˆ’iÎ¸Â }reduces to 2Â·cosÎ¸, while *e*^{iÎ¸Â }âˆ’Â *e*^{âˆ’iÎ¸Â }reduces to 2Â·*i*Â·sinÎ¸.]

Now, it’s easy to see that, if we’d have used +A rather thanÂ âˆ’A, we would have gotten something very similar:

- C
_{1}(t) = âŒ© 1 | Ïˆ âŒª = (1/2)Â·*e*^{âˆ’(i/Ä§)Â·(E0Â + A)Â·tÂ }+ (1/2)Â·*e*^{âˆ’(i/Ä§)Â·(E0Â âˆ’ A)Â·tÂ }=*e*^{âˆ’(i/Ä§)Â·E0Â·t}Â·cos[(A/Ä§)Â·t] - C
_{2}(t) = âŒ© 2 | Ïˆ âŒª = (1/2)Â·*e*^{âˆ’(i/Ä§)Â·(E0Â + A)Â·tÂ }â€“ (1/2)Â·*e*^{âˆ’(i/Ä§)Â·(E0Â âˆ’ A)Â·tÂ }=Â âˆ’*i*Â·*e*^{âˆ’(i/Ä§)Â·E0Â·t}Â·sin[(A/Ä§)Â·t]

So we get aÂ *minusÂ *sign in front of ourÂ C_{2}(t) function, because cos(Î±) = cos(â€“Î±) but sin(Î±) = âˆ’sin(Î±). However, the associated probabilities are exactly the same. For both, we get the sameÂ P_{1}(t) and P_{2}(t) functions:

- P
_{1}(t) =Â |C_{1}(t)|^{2}Â = cos^{2}[(A/Ä§)Â·t] - P
_{2}(t) =Â |C_{2}(t)|^{2Â }= sin^{2}[(A/Ä§)Â·t]

[Remember: theÂ *absoluteÂ *square of *iÂ *and âˆ’* iÂ *is |

*i*|

^{2Â }= +âˆš1

^{2 }= +1 andÂ

*|*âˆ’

*|*

*i*^{2}

*(âˆ’1)*

^{Â }=^{2}

*|*

*|**i*^{2Â }= +1 respectively, so the

*i*andÂ âˆ’

*i*Â in the two C

_{2}(t) formulas disappear.]

You’ll remember the graph:

Of course, you’ll say: that plus or minus sign in front of C_{2}(t) should matterÂ *somehow*, doesn’t it? Well… Think about it. Taking the absolute square of some complex number â€“ or some complexÂ *function ,Â *in this case!Â â€“ amounts to multiplying it with its complex conjugate. Because the complex conjugate of a product is the product of the complex conjugates, it’s easy to see what happens: theÂ *e*^{âˆ’(i/Ä§)Â·E0Â·t}Â factor in C_{1}(t) =Â *e*^{âˆ’(i/Ä§)Â·E0Â·t}Â·cos[(A/Ä§)Â·t] andÂ C_{2}(t) = Â±*i*Â·*e*^{âˆ’(i/Ä§)Â·E0Â·t}Â·sin[(A/Ä§)Â·t] gets multiplied by *e*^{+(i/Ä§)Â·E0Â·t}Â and, hence, doesn’t matter: *e*^{âˆ’(i/Ä§)Â·E0Â·t}Â·*e*^{+(i/Ä§)Â·E0Â·t}Â = *e*^{0Â }= 1. The cosine factor inÂ C_{1}(t) =Â *e*^{âˆ’(i/Ä§)Â·E0Â·t}Â·cos[(A/Ä§)Â·t] isÂ *real*, and so its complex conjugate is the same. Now, theÂ Â±*i*Â·sin[(A/Ä§)Â·t] factor in C_{2}(t) = Â±*i*Â·*e*^{âˆ’(i/Ä§)Â·E0Â·t}Â·sin[(A/Ä§)Â·t] is a *pure* imaginary number, and so its complex conjugate is its opposite. For some reason, we’ll find similar solutions for all of the situations we’ll describe below: the factor determining the *probabilityÂ *will either be real or, else,Â a *pure* imaginary number. Hence, from a math point of view, it really doesn’t matter if we take +A or âˆ’A for Â or Â *real* factor for those H_{12Â }and H_{21}Â coefficients. We just need to be consistent in our choice, and I must assume that, in order to be consistent, Feynman likes to think of our nitrogen atom *borrowing *some energy from the system and, hence, * temporarilyÂ reducing*Â its energy by an amount that’s equal to âˆ’A.

*If you have a better interpretation, please do let me know!*ðŸ™‚

OK. We’re done with this section… Except… Well… I have to show you how we got those C_{1}(t) and C_{1}(t) functions, no? Let me copy Feynman here:

Note that the ‘trick’ involving the addition and subtraction of the differential equations is a trick we’ll use quite often, so please do have a look at it. As for the value of the *a* and *b* coefficients â€“ which, as you can see, we’ve equated to 1 in our solutions for C_{1}(t) and C_{1}(t) â€“ we get those because of the following starting condition: we assume that at t = 0, the molecule will be in state 1. Hence, we assumeÂ C_{1}(0) = 1 and C_{2}(0) = 0. In other words: we assume that we start out on that P_{1}(t) curve in that graph with the probability functions above, so the C_{1}(0) = 1 and C_{2}(0) = 0 starting condition is equivalent toÂ P_{1}(0) = 1 and P_{1}(0) = 0. Plugging that in gives us *a*/2 + *b*/2 = 1 and *a*/2 âˆ’ *b*/2 = 0, which is possible only ifÂ *a* = *b* = 1.

Of course, you’ll say: what if we’d choose to start out with state 2, so our starting condition isÂ P_{1}(0) = 0 and P_{1}(0) = 1? ThenÂ *a* = 1 andÂ *b* = âˆ’1, and we get the solution we got when equating H_{12Â }and H_{21}Â with +A, rather than with âˆ’A. So you can think about that symmetry once again: when we’re in free space, then it’s quite arbitrary what we call ‘up’ or ‘down’.

So… Well… That’s all great. I should, perhaps, just add one more note, and that’s on that A/Ä§Â *value*. We calculated it in the previous post, because we wanted to actually calculate theÂ *periodÂ *of thoseÂ P_{1}(t) and P_{2}(t) functions. Because we’re talking the *squareÂ *of a cosine and a sine respectively, the period is equal to Ï€, rather than 2Ï€, so we wrote:Â (A/Ä§)Â·T = Ï€ â‡” T = Ï€Â·Ä§/A. Now, theÂ *separationÂ *between the two energy levels E_{0Â }+ A and E_{0Â }âˆ’ A, so that’s 2A, has been *measuredÂ *as being equal, more or less, to 2A â‰ˆ 10^{âˆ’4Â }eV.

How does one *measureÂ *that? As mentioned above, I’ll show you, in a moment, that, when applying some external field, the plus and minus sign *doÂ *matter, and theÂ *separationÂ *between those two energy levels E_{0Â }+ A and E_{0Â }âˆ’ A will effectively represent something physical. More in particular, we’ll have *transitions* from one energy level to another and that corresponds to electromagnetic radiation being emitted or absorbed, and so there’s a relation between the energy and the *frequencyÂ *of that radiation. To be precise, we can writeÂ 2A =Â hÂ·f_{0}. The frequency of the radiation that’s being absorbed or emitted is 23.79 GHz, which corresponds to microwave radiation with a wavelength ofÂ Î» = *c*/f_{0}Â =Â 1.26 cm. Hence,Â 2Â·A â‰ˆ 25Ã—10^{9}Â Hz *times* 4Ã—10^{âˆ’15} eVÂ·s = 10^{âˆ’4Â }eV, indeed, and, therefore, we can write: T = Ï€Â·Ä§/A â‰ˆ 3.14 Ã— 6.6Ã—10^{âˆ’16Â }eVÂ·s *divided by* 0.5Ã—10^{âˆ’4Â }eV, so that’s 40Ã—10^{âˆ’12Â }seconds = 40 *pico*seconds. That’s 40Â *trillionthsÂ *of a seconds. So that’sÂ *veryÂ *short, and surely *muchÂ *shorter than the time that’s associated with, say, a freely emitting sodium atom, which is of the order of 3.2Ã—10^{âˆ’8Â }seconds. You may think that makes sense, because theÂ *photonÂ *energy is so much lower: a sodium light photon is associated with an energy equal to E = hÂ·*f* = 500Ã—10^{12}Â Hz *times* 4Ã—10^{âˆ’15Â } eVÂ·s = 2 eV, so that’s *20,000* times 10^{âˆ’4Â }eV.

There’s a funny thing, however. An oscillation of a frequency of 500 *tera*-hertzÂ that lastsÂ 3.2Ã—10^{âˆ’8Â }seconds is equivalent to 500Ã—10^{12}Â Hz *times* 3.2Ã—10^{âˆ’8Â }sÂ â‰ˆ 16 million cycles. However, an oscillation of a frequency of 23.97 *giga*-hertzÂ that only lasts 40Ã—10^{âˆ’12Â }seconds is equivalent to 23.97Ã—10^{9}Â Hz *times* 40Ã—10^{âˆ’12Â }sÂ â‰ˆ 1000Ã—10^{âˆ’3 }= 1 ! *One cycle only?Â *We’re surelyÂ *notÂ *talking resonance here!

So… Well… I am just flagging it here. We’ll have to do some more thinking about that later. [I’ve added an addendum that may or may not help us in this regard. :-)]

**The two-state system in a field**

As mentioned above, when there is no external force field, we define the ‘up’ or ‘down’ direction of the nitrogen atom was defined with regard to its its spin around its axis of symmetry, so with regard to the molecule itself. However, when we apply an external *electromagneticÂ *field, as shown below, we do have some external reference frame.

Now, the external reference frame â€“ i.e. the *physicsÂ *of the situation, reallyÂ â€“Â may make it more convenient to define the whole system using another set of base states, which we’ll refer to as I and II, rather than 1 and 2. Indeed, you’ve seen the picture below: it shows aÂ *state selector*, or a filter as we called it. In this case, there’s a filtering according to whether our ammonia molecule is in state I or, alternatively, state II. It’s like a Stern-Gerlach apparatus splitting an electron beam according to theÂ *spinÂ *state of the electrons, which is ‘up’ or ‘down’ too, but in a totally different way than our ammonia molecule. Indeed, the ‘up’ and ‘down’ spin of an electron has to do with itsÂ *magneticÂ *moment and itsÂ *angular momentum*. However, there are a lot of similarities here, and so you may want to compare the two situations indeed, i.e. the electron beam in an inhomogeneous magnetic field versus the ammonia beam in an inhomogeneous electric field.

Now, when reading Feynman, as he walks us through the relevant *LectureÂ *on all of this, you get the impression that it’s the I and II states only that have some kind ofÂ *physicalÂ *or *geometricÂ *interpretation. That’s not the case. Of course, the diagram of the state selector above makes itÂ *veryÂ *obvious that theseÂ new I and II base states make *very much*Â sense in regard to the orientation of the field, i.e. with regard to *external space*,Â rather than with respect to the position of our nitrogen atom vis-Ã¡-vis the hydrogens. But… Well… Look at the image below: the *direction* of the field (which we denote by Îµ because we’ve been using the E for energy) obviously matters when defining the old ‘up’ and ‘down’ states of our nitrogen atom too!

In other words, our previousÂ | 1 âŒª and | 2 âŒª base states acquire a new meaning too: it obviously matters whether or not the *electric dipole momentÂ *of the molecule is in the *same* or, conversely, in the *opposite* direction of the field. To be precise, the presence of the electromagnetic field suddenly gives theÂ energy levels that we’d associate with these two states a veryÂ different ** physical** interpretation.

Indeed, from the illustration above, it’s easy to see that the electric dipole moment of this particular molecule in state 1 is in the opposite direction and, therefore, temporarily ignoring the amplitude to flip over (so we do *notÂ *think ofÂ A for just a brief little moment), **the energy that we’d associate with state 1 would be equal to E _{0Â }+ Î¼Îµ. Likewise, the energy we’d associate with state 2 is equal to E_{0Â }âˆ’ Î¼Îµ.** Â Indeed, you’ll remember that the (potential) energy of an electric dipole is equal to the vector dot product of the electric dipole moment

**Î¼**and the field vector

**Îµ**, but with a minus sign in front so as to get the sign for the energy righ. So the energy isÂ equal to âˆ’

**Î¼**Â·

**Îµ**Â = âˆ’|

**Î¼**|Â·|

**Îµ**|Â·cosÎ¸, with Î¸ the angle between both vectors. Now, the illustration above makes it clear that state 1 and 2 are defined for Î¸ = Ï€ and Î¸ = 0 respectively. [And, yes! Please

*do*note that state 1 is theÂ

*highestÂ*energy level, because it’s associated with the highest potential energy: the electric dipole moment

**Î¼**Â of our ammonia moleculeÂ

*willÂ*â€“ obviously!Â â€“ want to align itself with the electric field

**Îµ**Â ! Just think of what it would imply to

*turn*the molecule in the field!]

Therefore, using the sameÂ *hunchesÂ *as the ones we used in the *free space* example, Feynman suggests that, when some external electric field is involved, we should use the following Hamiltonian matrix:

So we’ll need to solve a similar set of differential equations with *this* Hamiltonian now. We’ll do that later and, as mentioned above, it will be more convenient to switch to another set of base states, or another ‘representation’ as it’s referred to. But… Well… Let’s not get too much ahead of ourselves: I’ll say something about that before we’ll start solving the thing, but let’s first look at that Hamiltonian once more.

When I say that Feynman uses the same clues here, then… Well.. That’s true and not true. You should note that the diagonal elements in the Hamiltonian above areÂ *notÂ *the same: E_{0Â }+ Î¼ÎµÂ â‰ Â E_{0Â }+ Î¼Îµ. So we’ve lost that symmetry of free space which, from a math point of view, was reflected in those identicalÂ H_{11 }= H_{22Â }= E_{0Â }coefficients.

That should be obvious from what I write above: state 1 and state 2 are no longer those 1 and 2 states we described when looking at the molecule in free space. Indeed, theÂ | 1 âŒª and | 2 âŒª states are still ‘up’ or ‘down’, but the illustration above also makes it clear we’re defining state 1 and state 2 not only with respect to the molecule’s spin around its own axis of symmetry but also vis-Ã¡-vis some direction in space. To be precise, we’re defining state 1 and state 2 here with respect to the direction of the electric field **Îµ**. NowÂ *thatÂ *makes aÂ really big difference in terms of interpreting what’s going on.

In fact, the ‘splitting’ of the energy levels because of that amplitude A is nowÂ something *physical *too, i.e.Â something that goesÂ *beyond*Â just modeling the uncertainty involved. In fact, we’ll find it convenient to distinguish two new energy levels, which we’ll write asÂ E_{IÂ }=Â E_{0Â }+ A and E_{IIÂ }=Â E_{0Â }âˆ’ A respectively. They are, of course, related to those new base states | I âŒª and | II âŒª that we’ll want to use. So the E_{0Â }+ A and E_{0Â }âˆ’ A energy levels themselves will acquire some physical meaning, and especially the *separationÂ *between them, i.e. the value of 2A. Indeed,Â E_{IÂ }=Â E_{0Â }+ A and E_{IIÂ }=Â E_{0Â }âˆ’ A will effectively represent an ‘upper’ and a ‘lower’ energy level respectively.

But, again, I am getting ahead of myself.Â Let’s first, as part of working towards a solution for our equations, look at what happens if and when we’d switch to another representation indeed.

**Switching to another representation**

Let me remind you of what I wrote in my post on quantum mathÂ in this regard. The *actualÂ *state of our ammonia molecule â€“ or any quantum-mechanical system really â€“Â is always to be described in terms of a set of base states. For example, if we have two possible base states only, weâ€™ll write:

| Ï† âŒª = | 1 âŒª C_{1}Â +Â | 2 âŒª C_{2}

You’ll say: why? Our molecule is obviously always in *either* state 1Â *orÂ *state 2, isn’t it? Well… Yes and no. That’s the mystery of quantum mechanics: it is and it isn’t. As long as we don’t measure it, there is anÂ *amplitudeÂ *for it to be in state 1 and an amplitude for it to be in state 2. So we can only make sense of its state by actually calculating âŒ© 1Â | Ï† âŒª and âŒ© 2 |Â Ï† âŒª which, unsurprisingly are equal to âŒ© 1Â | Ï† âŒª = âŒ© 1 | 1 âŒª C_{1}Â +Â âŒ© 1 | 2 âŒª C_{2}Â = C_{1}(t)Â and âŒ© 2 | Ï† âŒª = âŒ© 2 | 1 âŒª C_{1}Â +Â âŒ© 2 | 2 âŒª C_{2}Â = C_{2}(t) respectively, and so these two functions give us the probabilities P_{1}(t) and Â P_{2}(t) respectively. So that’s SchrÃ¶dinger’s cat really: the cat is dead *orÂ *alive, but we don’t know until we open the box, and we only have a *probability function* â€“ so we can say that it’s *probably *dead orÂ *probably *alive, depending on the odds*Â â€“ *as long as we doÂ *notÂ *open the box. It’s as simple as that.

Now, the ‘dead’ and ‘alive’ condition are, obviously, the ‘base states’ in SchrÃ¶dinger’s rather famous example, and we can write them as | *DEAD*Â âŒª and | *ALIVE*Â âŒªÂ you’d agree it would be difficult to find another representation. For example, it doesn’t make much sense to say that we’ve rotated the two base states over 90 degrees and we now have two new states equal to (1/âˆš2)Â·| *DEAD*Â âŒª â€“ (1/âˆš2)Â·| *ALIVE*Â âŒª and (1/âˆš2)Â·| *DEAD*Â âŒª + (1/âˆš2)Â·| *ALIVE*Â âŒª respectively. There’s no *directionÂ *in space in regard to which we’re defining those two base states: dead is dead, and alive is alive.

The situation really resembles our ammonia molecule *in free space*: there’s no external reference against which to define the base states. However, as soon as some external field is involved, we *doÂ *have a direction in space and, as mentioned above, our base states are now definedÂ *with respect to a particular orientation in space*. That implies two things. The first is that we should no longer say that our molecule will always be in *either* state 1Â *orÂ *state 2. There’s no reason for it to be perfectly aligned with or against the field. Its orientation can be anything really, and so its state is likely to be some combination of those two *pureÂ *base states | 1 âŒª and | 2 âŒª.

The second thing is that we may choose* another set of base states*, and specify the very same state *in terms of the new base states*. So, assuming we choose some other set of base statesÂ | I âŒª and | II âŒª, we can write the very same stateÂ | Ï† âŒª = | 1 âŒª C_{1}Â +Â | 2 âŒª C_{2Â }as:

| Ï† âŒª = | I âŒª C_{I}Â +Â | II âŒª C_{II}

It’s really like what you learned about vectors in high school: one can go from one set of base *vectors* to another by aÂ *transformation,Â *such as, for example, aÂ *rotation*, or a *translation*. It’s just that, just like in high school, we need some direction in regard to which we *define*Â our rotation or our translation.

ForÂ *stateÂ *vectors, I showed how aÂ *rotation *of base statesÂ worked in one of my posts on two-state systems. To be specific, we had the following relation between the two representations:

The (1/âˆš2) factor is there because of the normalization condition, and theÂ two-by-two matrix equals the transformation matrix for a* rotation* of a state filtering apparatus *about the y-axis*, over an angle equal to (minus)Â 90 degrees, which we wrote as:

The *y*-axis? *What y-axis?Â **What state filtering apparatus**?*Â Just relax. Think about what you’ve learned already. The orientations are shown below: the S apparatus separates ‘up’ and ‘down’ states along the z-axis, while the T-apparatus does so along an axis that is tilted,Â *about the y-axis*, over an angle equal toÂ Î±, orÂ Ï†, as it’s written in the table above.

Of course, we don’t *really* introduce an apparatus at this or that angle. We just introduced an electromagnetic field, which re-defined our | 1 âŒª and | 2 âŒª base states and, therefore, through the rotationalÂ *transformation matrix*,Â also *definesÂ *ourÂ | I âŒª andÂ | II âŒª base states.

[…] You may have lost me by now, and so then ** you’ll want to skip to the next section**. That’s fine.

**Just remember that theÂ**. Having said that, if you’re reading this post, and you want to understand it,

*representationsÂ*in terms of | I âŒª and | II âŒª base states or in terms of | 1 âŒª and | 2 âŒª base states are*mathematically*equivalent*truly*Â (because you want to

*truly*understand quantum mechanics), then you should try to stick with me here. ðŸ™‚Â Indeed, there’s a zillion things you could think about right now, but you should stick to the math now. Using thatÂ

*transformationÂ*matrix, we can relate the C

_{IÂ }and C

_{IIÂ }coefficients in the | Ï† âŒª = | I âŒª C

_{I}Â +Â | II âŒª C

_{II}Â expression to the C

_{IÂ }and C

_{IIÂ }coefficients in theÂ | Ï† âŒª = | 1 âŒª C

_{1}Â +Â | 2 âŒª C

_{2}Â expression. Indeed, we wrote:

- C
_{IÂ }=Â âŒ© I | Ïˆ âŒª = (1/âˆš2)Â·(C_{1 }âˆ’ C_{2}) - C
_{IIÂ }=Â âŒ© II | Ïˆ âŒª = (1/âˆš2)Â·(C_{1 }+ C_{2})

That’s exactly the same as writing:

OK. […]Â ** Waw!**Â You just took a huge leap, because we can now

*compare*the two sets of differential equations:

They’re *mathematically *equivalent, but theÂ mathematical* behavior*Â of the functions involved is very different.Â Indeed, unlike the C_{1}(t) and C_{2}(t) amplitudes, we find that the C_{I}(t) and C_{II}(t) amplitudes are stationary, i.e. the associated probabilities â€“ which we find by taking the absolute square of the amplitudes, as usualÂ â€“Â doÂ *notÂ *vary in time.Â To be precise, if you write it all out and simplify, you’ll find that theÂ C_{I}(t) and C_{II}(t) amplitudes are equal to:

- C
_{I}(t) = âŒ© I | Ïˆ âŒª = (1/âˆš2)Â·(C_{1 }âˆ’ C_{2}) =Â (1/âˆš2)Â·*e*^{âˆ’(i/Ä§)Â·(E0+ A)Â·t}Â = (1/âˆš2)Â·*e*^{âˆ’(i/Ä§)Â·EIÂ·t} - C
_{II}(t) = âŒ© II | Ïˆ âŒª = (1/âˆš2)Â·(C_{1 }+ C_{2}) =Â (1/âˆš2)Â·*e*^{âˆ’(i/Ä§)Â·(E0âˆ’ A)Â·tÂ }= (1/âˆš2)Â·*e*^{âˆ’(i/Ä§)Â·EIIÂ·t}

As the absolute square of the exponential is equal to one, the associated probabilities, i.e. |C_{I}(t)|^{2}Â and |C_{II}(t)|^{2}, are, quite simply, equal toÂ |1/âˆš2|^{2}Â = 1/2. Now, it is very tempting to say that this means that our ammonia molecule has an equal chance to be in state I or state II. In fact, while I may have said something like that in my previous posts, that’s *not* how one should interpret this. **The chance of our molecule beingÂ exactlyÂ in state I or state II, or in state 1 or state 2Â is varying with time, with the probability being ‘dumped’ from one state to the other all of the time.**

I mean… The electric dipole moment can point in any direction, really. So saying that our molecule has a 50/50 chance of being in state 1 or state 2 makes no sense. Likewise, saying that our molecule has a 50/50 chance of being in state I or state II makes no sense either. Indeed, the state of our molecule is specified by the | Ï† âŒª = | I âŒª C_{I}Â +Â | II âŒª C_{IIÂ }=Â | 1 âŒª C_{1}Â +Â | 2 âŒª C_{2Â }equations, and neither of these two expressions is a *stationaryÂ *state. They mix two frequencies, because they mix two energy levels.

Having said that,Â *we’re talking quantum mechanics here*Â and, therefore, an external inhomogeneous electric field will effectively split the ammonia molecules according to their state. The situation is really like what a Stern-Gerlach apparatus does to a beam of electrons: it will split the beam according to the electron’s spin, which is *eitherÂ *‘up’ or*, else*, ‘down’, as shown in the graph below:

The graph for our ammonia molecule, shown below, is very similar. The vertical axis measures the same: energy. And the horizontal axis measures Î¼Îµ, which increases with the strength of the electric field Îµ. So we see a similar ‘splitting’ of the energy of the molecule in an external electric field.

How should we explain this? It is *very *tempting to think that the presence of an external force field causes the electrons, or the ammonia molecule, to ‘snap into’ one of the two possible states, which are referred to as state I and state II respectively in the illustration of the ammonia state selector below. But… Well… Here we’re entering the murky waters of actually *interpretingÂ *quantum mechanics, for which (*a*) we have no time, and (*b*) we are not qualified. So you should just believe, or take for granted, what’s being shown here: an inhomogeneous electric field will split our ammonia beam according to their state, which we define as I and II respectively, and which are associated with the energy E_{0}+ A and E_{0}âˆ’ A Â respectively.

As mentioned above, you should note that these two states areÂ *stationary*. The Hamiltonian equations which, as they always do, describe theÂ *dynamicsÂ *of this system, imply that the amplitude to go from state I to state II, or vice versa, is zero. To make sure you ‘get’ that, I reproduce the associated Hamiltonian matrix once again:

Of course, that will change when we start our analysis of what’s happening in the maser. Indeed, we *will* have some *non*-zero H_{I,II}Â and H_{II,IÂ }amplitudes in the *resonantÂ *cavity of our ammoniaÂ *maser*, in which we’ll have an *oscillatingÂ *electric field and, as a result,Â *inducedÂ *transitions from state I to II and vice versa. However, that’s for later. While I’ll quickly insert the full picture diagram below, you should, for the moment, just think about those two stationary states and those two zeroes. ðŸ™‚

Capito? If not… Well… Start reading this post again, I’d say. ðŸ™‚

**Intermezzo: on approximations**

**Intermezzo: on approximations**

At this point, I need to say a few things about all of the approximations involved, because it can be quite confusing indeed. So let’s take a closer look at those energy levels and the related Hamiltonian coefficients.Â In fact, in his *Lectures*,Â Feynman shows us that we can always have a *general*Â solution for the Hamiltonian equations describing a two-state system whenever we have *constant *Hamiltonian coefficients. That general solution â€“ which, mind you, is derived assuming Hamiltonian coefficients that do *notÂ *depend on timeÂ â€“Â canÂ *alwaysÂ *be written in terms of two stationary base states, i.e. states with aÂ *definiteÂ *energy and, hence, aÂ *constantÂ *probability. The equations, and the two definite energy levels are:

That yields the following values for the energy levels for the *stationary* states:

Now, that’s *veryÂ *different from the E_{IÂ }= E_{0}+ A and E_{IIÂ }=Â E_{0}âˆ’ A energy levels for those stationary states we had defined in the previous section: *thoseÂ *stationary states had no square root, and noÂ Î¼^{2}Îµ^{2}, in their energy. In fact, that sort of answers the question: if there’s no external field, then thatÂ Î¼^{2}Îµ^{2}Â factor is zero, and the square root in the expression becomesÂ Â±âˆšA^{2Â }= Â±A. So then we’re back to ourÂ E_{IÂ }= E_{0}+ A and E_{IIÂ }=Â E_{0}âˆ’ A formulas.Â The whole point, however, is that **we willÂ actually have an electric field** in that cavity. Moreover, it’s going to be a field

**that varies in time**, which we’ll write:

Now, part of the confusion in Feynman’s approach is that he constantly switches between representing the system in terms of the I and II base states and the 1 and 2 base states respectively. For a good understanding, we should compare with our original representation of the dynamics in free space, for which the Hamiltonian was the following one:

*That* matrix can easily be related to the new one we’re going to have to solve, which is equal to:

The interpretation is easy if we look at that illustration again:

If the direction of the electric dipole moment is *oppositeÂ *to the directionÂ Îµ, then the associated energy is equal toÂ âˆ’**Î¼**Â·**Îµ**Â = âˆ’**Î¼**Â·**Îµ**Â = âˆ’|**Î¼**|Â·|**Îµ**|Â·cosÎ¸ =Â âˆ’Î¼Â·ÎµÂ·cos(Ï€) = +Î¼Îµ. Conversely, for state 2, we find âˆ’Î¼Â·ÎµÂ·cos(0) = âˆ’Î¼Îµ for the energy that’s associated with the dipole moment. You can and should think about the physics involved here, because they make sense! Thinking of amplitudes, you should note that theÂ +Î¼Îµ and âˆ’Î¼ÎµÂ terms effectively *change* the H_{11}Â and H_{22}Â coefficients, so they *changeÂ *the amplitude toÂ stay in state 1 or state 2 respectively. That, of course, will have an impact on the associated probabilities, and so that’s why we’re talking of *inducedÂ *transitions now.

Having said that, the Hamiltonian matrix above keeps theÂ âˆ’A forÂ H_{12}Â and H_{21}, so the matrix captures spontaneous transitions too!

Still… You may wonder why Feynman doesn’t use those E_{IÂ }and E_{IIÂ }formulas with the square root because that would give us some *exact *solution, wouldn’t it? The answer to that question is: maybe it would, but would you know how to solve those equations? We’ll have a *varyingÂ *field, remember? So our Hamiltonian H_{11}Â and H_{22}Â coefficients will no longer be constant, but *time-dependent*. As you’re going to see, it takes Feynman three pages to solve the whole thing using the +Î¼Îµ and âˆ’Î¼Îµ approximation. So just imagine how complicated it would be using that square root expression!Â [By the way, do have a look at those asymptotic curves in that illustration showing the splitting of energy levels above, so you see how that approximation looks like.]

So that’s the real answer: we need to simplify somehow, so as to get any solutions at all!

Of course, it’s all quite confusing because, after Feynman first notes that,Â for strong fields, the A^{2}Â in that square root is small as compared to Î¼^{2}Îµ^{2}, thereby justifying the use of the simplified E_{IÂ }= E_{0}+ Î¼Îµ = H_{11}Â andÂ E_{IIÂ }= E_{0}âˆ’ Î¼ÎµÂ = H_{22}Â coefficients, he continues and bluntly uses the very same square root expression to explain how that state selector works, saying that the electric field *in the state selector* will be rather weak and, hence, that Î¼Îµ will be much smaller than A, so one can use the following approximation for the square root in the expressions above:

The energy expressions then reduce to:

And then we can calculate the force on the molecules as:

So the electric field in the state selector is weak, but the electric field in the cavity is supposed to be strong, and so… Well… That’s it, really. The bottom line is that we’ve a beam of ammonia molecules thatÂ *are all in state I*, and it’s what happens withÂ *thatÂ *beam then, that is being described by our new set ofÂ differential equations:

**Solving the equations**

As all molecules in our ammonia beam are described in terms of theÂ | I âŒª andÂ | II âŒªÂ baseÂ states â€“ as evidenced by the fact that we say all molecules that enter the cavity are state I â€“ we need to switch to that representation. We do that by using that transformation above, so we write:

- C
_{IÂ }=Â âŒ© I | Ïˆ âŒª = (1/âˆš2)Â·(C_{1 }âˆ’ C_{2}) - C
_{IIÂ }=Â âŒ© II | Ïˆ âŒª = (1/âˆš2)Â·(C_{1 }+ C_{2})

Keeping these ‘definitions’ of C_{IÂ }and C_{IIÂ }in mind, you should then *addÂ *the two differential equations, divide the result by the square root of 2, and you should get the followingÂ *newÂ *equation:

*Please! Do it and verify the result! You want to learn something here, no?Â *ðŸ™‚

Likewise,Â *subtractingÂ *the two differential equations, we get:

Now, the problem is that the Hamiltonian constants here are *not *constant. To be precise, the electric field Îµ varies in time. We wrote:

So H_{I,II Â }and H_{II,I}, which are equal to Î¼Îµ, are *not constant*: we’ve got ** Hamiltonian coefficients that are a function of time themselves. **[…]

**So… Well… We just need to get on with it and try to finally**

*Â***this thing. Let me just copy Feynman as he grinds through this:**

*solve*This is only the first step in the process. Feynman just takes twoÂ *trialÂ *functions, which are really similar to the very general C_{1Â }=Â *a*Â·e^{â€“(i/Ä§)Â·H11Â·tÂ }function we presented when only one equation was involved, or â€“ if you prefer a set of two equations â€“ thoseÂ C_{I}(t) = *a*Â·*e*^{âˆ’(i/Ä§)Â·EIÂ·tÂ }and C_{I}(t) = *b*Â·*e*^{âˆ’(i/Ä§)Â·EIIÂ·}^{tÂ }equations above. The difference is that theÂ *coefficientsÂ *in front, i.e. Î³_{I}Â and Î³_{II}Â are *not *someÂ (complex) constant, but functions of time themselves. The next step in the derivation is as follows:

One needs to do a bit of gymnastics here as well to follow what’s going on, but please do check and you’ll see it works. Feynman derives *another* set of differential equations here, and they specify these Î³_{I}Â = Î³_{I}(t) and Î³_{II}Â = Î³_{II}(t) functions. These equations are written in terms of the frequency of the field, i.e. Ï‰, and the *resonant *frequency Ï‰_{0}, which we mentioned above when calculating that 23.79 GHz frequency from theÂ 2A =Â hÂ·f_{0}Â equation. SoÂ Ï‰_{0}Â is the same molecular resonance frequency but expressed as an *angularÂ *frequency, soÂ Ï‰_{0}Â =Â f_{0}/2Ï€ =Â Ä§/2A. He then proceeds to simplify, using assumptions one should check. He thenÂ continues:

That gives us what we presented in the previous post:

So… Well… What to say? I explained those probability functions in my previous post, indeed. We’ve gotÂ two probabilities here:

- P
_{IÂ }=Â cos^{2}[(Î¼Îµ_{0}/Ä§)Â·t] - P
_{IIÂ }=Â sin^{2}[(Î¼Îµ_{0}/Ä§)Â·t]

So thatâ€™s just like the P_{1Â }=Â Â cos^{2}[(A/Ä§)Â·t] and P_{2Â }=Â sin^{2}[(A/Ä§)Â·t] probabilities we found forÂ *spontaneousÂ *transitions. But so hereÂ we are talking *induced*Â transitions.

As you can see, the frequency and, hence, the period, depend on the *strength*,*Â *orÂ *magnitude*, of the electric field, i.e. the Îµ_{0Â }constant in the Îµ = 2Îµ_{0}cos(Ï‰Â·t) expression. The natural unit for measuring time would be the period once again, which we can easily calculate as (Î¼Îµ_{0}/Ä§)Â·T = Ï€ â‡” T = Ï€Â·Ä§/Î¼Îµ_{0}.

Now, we had thatÂ T =Â (Ï€Â·Ä§)/(2A) expression above, which allowed us to calculate theÂ *periodÂ *of the spontaneous transition frequency, which we found was like 40 *pico*seconds, i.e. 40Ã—10^{âˆ’12Â }seconds. Now, theÂ T =Â (Ï€Â·Ä§)/(2Î¼Îµ_{0}) is very similar, it allows us to calculate the *expected, average, or mean time *for an *induced *transition. In fact, if we write T* _{induced}*Â = (Ï€Â·Ä§)/(2Î¼Îµ

_{0}) and T

*Â =Â (Ï€Â·Ä§)/(2A), then we can take ratio to find:*

_{spontaneous}T* _{induced}*/T

*Â = [(Ï€Â·Ä§)/(2Î¼Îµ*

_{spontaneous}_{0})]/[(Ï€Â·Ä§)/(2A)] = A/Î¼Îµ

_{0}

ThisÂ A/Î¼Îµ_{0 }ratio is greater than one, so T* _{induced}*/T

*Â is greater than one, which, in turn, means that the presence of our electric field â€“ which, let me remind you,Â*

_{spontaneous}*dances*to the beat of the resonant frequency â€“ causes a

*slowerÂ*transition than we would have had if the oscillating electric field were

*notÂ*present.

**But – Hey!Â – that’s the wrong comparison!**Â Remember all molecules enter in aÂ

*stationary*

*state*, as they’ve been selected so as to ensure they’re in state I. So there is no such thing as a spontaneous transition frequency here! They’re all polarized, so to speak, and they would remain that way if there was no field in the cavity. So

*if there was no oscillating electric field, they would never transition*.

**Nothing would happen!**Well… In terms of our particular set of base states, of course! Why? Well… Look at the Hamiltonian coefficients H

_{I,IIÂ }= H

_{II,IÂ }= Î¼Îµ: these coefficients are zero if Îµ is zero. So… Well… That says it all.

SoÂ *that*‘s what it’s all about: *induced* emission and, as I explained in my previous post, because all molecules enter in state I, i.e. theÂ *upperÂ *energy state, literally, they all ‘dump’ aÂ *netÂ *amount of energy equal toÂ 2A into the cavity at the occasion of their first transition. The molecules then keep dancing, of course, and so they absorb and emit the same amount as they go through the cavity, but… Well… We’ve got aÂ *netÂ *contribution here, which is not only enough toÂ *maintainÂ *the cavity oscillations, but actually also provides a smallÂ *excessÂ *of power that can be drawn from the cavity as microwave radiation of the same frequency.

As Feynman notes, anÂ *exactÂ *description of what actually happens requires an understanding of the quantum mechanics of the field in the cavity, i.e.Â *quantum field theory*, which I haven’t studied yet. But… Well… That’s for later, I guess. ðŸ™‚

**Post scriptum**: The sheer length of this post shows we’re not doing something that’s easy here. Frankly, I feel the whole analysis is still quite obscure, in the sense thatÂ â€“ despite looking at this thing again and againÂ â€“ it’s hard to sort of interpret what’s going on,Â *in a physical sense that is*. But perhaps one shouldn’t try that. I’ve quoted Feynman’s view on how easy or how difficult it is to ‘understand’ quantum mechanics a couple of times already, so let me do it once more:

“Because atomic behavior is so unlike ordinary experience, it is very difficult to get used to, and it appears peculiar and mysterious to everyoneâ€”both to the novice and to the experienced physicist. Even the experts do not understand it the way they would like to, and it is perfectly reasonable that they should not, because all of direct, human experience and human intuition applies to large objects.”

So… Well… I’ll grind through the remaining Lectures nowÂ â€“ I am halfway through Volume III nowÂ â€“ and then re-visit all of this. Despite Feynman’s warning, I want to understand it the way I like to, even if I don’t quite know what way that is right now. ðŸ™‚

**Addendum**: As for those cycles and periods, I noted a couple of times already that the Planck-Einstein equation E = hÂ·*f*Â Â can usefully be re-written as E/*fÂ *= h, as it gives aÂ *physicalÂ *interpretation to the value of the Planck constant. In fact, I said h is the energy that’s associated withÂ *oneÂ *cycle, regardless of the frequency of the radiation involved. Indeed, the energy of a photon divided by the number of cycles per second, should give us the energyÂ *per cycle*, no?

Well… Yes and no. Planck’s constant h and the frequencyÂ *fÂ *are both expressed referencing the time unit. However, if we say that a sodium atom emits *oneÂ *photon only as its electron transitions from a higher energy level to a lower one, and if we say that involves a decay time of the order ofÂ 3.2Ã—10^{âˆ’8Â }seconds, then what we’re saying really is that a sodium light photon will ‘pack’ like 16 million cycles, which is what we get when we *multiply*Â the number of cycles *per second*Â (i.e. the mentioned frequency of 500Ã—10^{12 }Hz) by the decay time (i.e.Â 3.2Ã—10^{âˆ’8Â }seconds): (500Ã—10^{12 }Hz)Â·(3.2Ã—10^{âˆ’8Â }s) = 16 Ã—10^{6Â }cycles, indeed. So the *energyÂ *per cycle is 2.068 eVÂ (i.e. the photon energy) divided by 16Ã—10^{6}, so that’s 0.129Ã—10^{âˆ’6Â }eV. Unsurprisingly, that’s what we get when we we divide h by 3.2Ã—10^{âˆ’8Â }s: (4.13567Ã—10^{âˆ’15})/(3.2Ã—10^{âˆ’8Â }s) = 1.29Ã—10^{âˆ’7Â }eV. We’re just putting some values in to the E/(*fÂ·*T)*Â *= h/T equation here.

The logic for that 2A =Â hÂ·f_{0}Â is the same. TheÂ frequency of the radiation that’s being absorbed or emitted is 23.79 GHz, so the photon energy is (23.97Ã—10^{9}Â Hz)Â·(4.13567Ã—10^{âˆ’15}Â eVÂ·s) â‰ˆÂ 1Ã—10^{âˆ’4Â }eV. Now, we calculated the transition periodÂ T as T = Ï€Â·Ä§/A â‰ˆ (Ï€Â·6.626Ã—10^{âˆ’16Â }eVÂ·s)/(0.5Ã—10^{âˆ’4Â }eV) â‰ˆ 41.6Ã—10^{âˆ’12Â }seconds. Now,Â an oscillation of a frequency of 23.97 *giga*-hertzÂ that only lasts 41.6Ã—10^{âˆ’12Â }seconds is an oscillation of *one cycleÂ *only. The consequence is that, when we continue this style of reasoning, we’d have a photon that packs all of its energy into one cycle!

Let’s think about what this implies in terms of the density in space. The wavelength of our microwave radiation is 1.25Ã—10^{âˆ’2Â }m, so we’ve got a ‘density’ ofÂ 1Ã—10^{âˆ’4Â }eV/1.25Ã—10^{âˆ’2Â }m = 0.8Ã—10^{âˆ’2 }eV/m = 0.008 eV/m. The wavelength of our sodium light is 0.6Ã—10^{âˆ’6Â }m, so we get a ‘density’ of 1.29Ã—10^{âˆ’7Â }eV/0.6Ã—10^{âˆ’6Â }m = 2.15Ã—10^{âˆ’1Â }eV/m = 0.215 eV/m. So the energy ‘density’ of our sodium light is 26.875 times that of our microwave radiation. ðŸ™‚

Frankly, I am not quite sure if calculations like this make much sense. In fact, when talking about energy densities, I should review my posts on the Poynting vector. However, they may help you think things through. ðŸ™‚

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://en.support.wordpress.com/copyright-and-the-dmca/

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:https://en.support.wordpress.com/copyright-and-the-dmca/

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:https://en.support.wordpress.com/copyright-and-the-dmca/

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:https://en.support.wordpress.com/copyright-and-the-dmca/

Some content on this page was disabled on June 17, 2020 as a result of a DMCA takedown notice from Michael A. Gottlieb, Rudolf Pfeiffer, and The California Institute of Technology. You can learn more about the DMCA here: