Re-visiting relativity and four-vectors: the proper time, the tensor and the four-force

Pre-script (dated 26 June 2020): Our ideas have evolved into a full-blown realistic (or classical) interpretation of all things quantum-mechanical. In addition, I note the dark force has amused himself by removing some material. So no use to read this. Read my recent papers instead. šŸ™‚

Original post:

My previous post explained how four-vectors transformĀ from one reference frame to the other. Indeed, a four-vector isĀ notĀ just some one-dimensional array of four numbers: it represent something—a physical vector that… Well… Transforms like a vector. šŸ™‚ So whatĀ vectorsĀ are we talking about? Let’s see what we have:

  1. We knew the position four-vector already, which we’ll write as xμ = (ct, x, y, z) = (ct, x).
  2. We also proved that Aμ = (Φ,Ā Ax, Ay, Az) = (Φ,Ā A)Ā is a four-vector: it’s referred to as the four-potential.
  3. We also know the momentumĀ four-vector from theĀ LecturesĀ on special relativity. We write it as pμ = (E, px, py, pz) = (E, p), with E = γm0, p = γm0v, and γ = (1āˆ’v2/c2)āˆ’1/2Ā or, forĀ cĀ = 1, γ =Ā (1āˆ’v2)āˆ’1/2

To show that it’s notĀ just a matter of adding some fourth t-component to aĀ three-vector, Feynman gives the example of the four-velocity vector. We have vxĀ = dx/dt,Ā vyĀ = dy/dt and vzĀ = dz/dt, but a vμ = (d(ct)/dt, dx/dt, dy/dt, dz/dt) = (c, dx/dt, dy/dt, dz/dt) ‘vector’ is, obviously, not a four-vector. [Why obviously? The inner productĀ vμvμ Ā is not invariant.] In fact,Ā Feynman ‘fixes’ the problem by noting that ct, x, y and z have the ‘right behavior’, but the d/dt operator doesn’t. The d/dt operator isĀ not an invariant operator. So how does he fix it then? He tries the (1āˆ’v2/c2)āˆ’1/2Ā·d/dt operator and, yes, it turns out we do get a four-vector then. In fact, we get that four-velocity vector uμ that we were looking for:four-velocity vector[Note we assume we’re using equivalentĀ time and distance units now, soĀ cĀ = 1 andĀ v/c reduces to a new variable v.]

Now how do we know this is four-vector? How can we prove this one? It’s simple. We can get it from ourĀ pμ = (E, p) by dividing it by m0, which is an invariantĀ scalar in four dimensions too. Now, it is easy to see that a division by an invariantĀ scalar doesĀ notĀ change the transformation properties. So just write it all out, and you’ll see that pμ/m0Ā = uμ and, hence, that uμ is a four-vector too. šŸ™‚

We’ve got an interesting thing here actually: division by an invariant scalar, or applying that (1āˆ’v2/c2)āˆ’1/2Ā·d/dt operator, which is referred to as an invariant operator, on a four-vector will give us another four-vector. Why is that? Let’s switch to compatible time and distance units soĀ c = 1 so to simplify the analysis that follows.

The invariant (1āˆ’v2)āˆ’1/2Ā·d/dt operatorĀ and the proper time s

Why is theĀ (1āˆ’v2)āˆ’1/2Ā·d/dt operator invariant? Why does it ‘fix’ things? Well… Think about the invariant spacetime interval (Ī”s)2Ā = Ī”t2Ā āˆ’ Ī”x2Ā āˆ’ Ī”y2Ā āˆ’Ā Ī”z2Ā going to the limit (ds)2Ā = dt2Ā āˆ’ dx2Ā āˆ’ dy2Ā āˆ’ dz2Ā . Of course, we can and should relate this to an invariant quantity s = ∫ ds. Just like Ī”s, this quantity also ‘mixes’ time and distance. Now, we could try to associate some derivative d/ds with it because, as Feynman puts it, “it should be a nice four-dimensional operation because it is invariant with respect to a Lorentz transformation.” Yes. It should be. So let’s relate ds to dt and see what we get. That’s easy enough: dx = vxĀ·dt,Ā dy = vyĀ·dt, dz = vzĀ·dt, so we write:

(ds)2Ā = dt2Ā āˆ’ vx2Ā·dt2Ā āˆ’ vy2Ā·dt2Ā āˆ’Ā vz2Ā·dt2 ⇔ (ds)2Ā = dt2Ā·(1 āˆ’ vx2Ā āˆ’ vy2Ā āˆ’Ā vz2) = dt2Ā·(1 āˆ’ v2)

and, therefore, ds = dtĀ·(1āˆ’v2)1/2. So our operator d/ds is equal to (1āˆ’v2)āˆ’1/2Ā·d/dt, and we can apply it toĀ anyĀ four-vector, as we are sure that, as an invariant operator, it’s going to give us another four-vector. I’ll highlight the result, because it’s important:

The d/ds = (1āˆ’v2)āˆ’1/2Ā·d/dt operator is an invariant operator for four-vectors.

For example, if we apply it to xμ = (t, x, y, z), we get the very same four-velocity vector μμ:

dxμ/ds = uμ = pμ/m0

Now, if you’re somewhat awake, you should ask yourself: what is this s, really,Ā and what is this operator all about? Our new function s = ∫ ds is notĀ the distance function, as it’s got both time and distance in it. Likewise, the invariant operatorĀ d/ds = (1āˆ’v2)āˆ’1/2Ā·d/dt has both time and distance in it (the distance is implicit in the v2Ā factor). Still, it is referred to as the proper timeĀ along the path of a particle. Now why is that? If it’s got distance andĀ time in it, why don’t we call it the ‘proper distance-time’ or something?

Well… The invariant quantity s actually is the time that would be measured by a clock that’s moving along, in spacetime, with the particle. Just think of it: in the reference frame of the moving particle itself, Ī”x, Ī”yĀ and Ī”zĀ must be zero, because it’s not moving in its own reference frame. So theĀ (Ī”s)2Ā = Ī”t2Ā āˆ’ Ī”x2Ā āˆ’ Ī”y2Ā āˆ’Ā Ī”z2Ā reduces to (Ī”s)2Ā = Ī”t2, and so we’re only adding time to s. Of course, this view of things implies that theĀ proper time itself is fixed only up to some arbitrary additive constant, namely the setting of the clock at some event along the ‘world line’ of our particle, which is its path in four-dimensional spacetime. But… Well… In a way, s is the ‘genuine’ or ‘proper’ time coming with the particle’s reference frame, and so that’s why Einstein called it like that. You’ll see (later) that it plays a very important role in general relativity theory (which is a topic we haven’t discussed yet: we’ve only touched special relativity, so no gravity effects).

OK. I know this is simple and complicated at the same time: the math is (fairly) easy but, yes, it may be difficult to ‘understand’ this in some kind of intuitiveĀ way. But let’s move on.

The four-force vector fμ

We know the relativistically correct equation for the motionĀ of some charge q. It’s just Newton’s Law F = dp/dt = d(mv)/dt. The only difference is that we areĀ not assuming that m is some constant. Instead, we use the pĀ = γm0vĀ formula to get:

motion

How can we get a four-vector for the force? It turns out that we get it when applying our new invariant operator to the momentum four-vector pμ = (E, p), so we write: fμ = dpμ/ds. But pμ = m0uμ = m0dxμ/ds, so we can re-write this as fμ = d(m0·dxμ/ds)/ds, which gives us a formula which is reminiscent of the Newtonian F = ma equation:

force formula

WhatĀ isĀ this thing? Well… It’s not so difficult to verify that the x, y and z-components are just our old-fashioned Fx,Ā FyĀ andĀ Fz, so these are the components of F. The t-component is (1āˆ’v2)āˆ’1/2Ā·dE/dt. Now, dE/dt is the time rate of change of energy and, hence, it’s equal to the rate of doing work on our charge, which is equal to F•v. So we can write fμ as:

froce

The force and the tensor

We will now derive that formula which we ended the previous postĀ with. We start with calculating the spacelike components of fμ from the Lorentz formula F = q(E + vƗB). [The terminology is nice, isn’t it? The spacelike components of the four-force vector!Ā Now thatĀ sounds impressive, doesn’t it? But so… Well… It’s really just the old stuff we know already.]Ā So we start with fxĀ =Ā Fx, and write it all out:

fx

What a monster! But,Ā hey! We can ‘simplify’ this by substituting stuff by (1) the t-, x-, y- and z-components of the four-velocity vector uμ and (2) the components of our tensor Fμν = [Fij] = [āˆ‡iAjĀ āˆ’Ā āˆ‡jAi] with i, j = t, x, y, z. We’ll also pop in the diagonal FxxĀ = 0 element, just to make sure it’s all there. We get:

fx 2

Looks better, doesn’t it? šŸ™‚ Of course, it’s just the same, really. This is just an exercise in symbolism. Let me insert the electromagnetic tensor we defined in our previous post, just as a reminder of what that Fμν matrix actually is:

electromagnetic tensor final

If you read my previous post, this matrix – or the concept of a tensor – has no secrets for you. Let me briefly summarize it, because it’s an important result as well. The tensor is (a generalization of) the cross-product inĀ four-dimensional space. We take two vectors:Ā aμ =Ā (at, ax, ay, az) and bμ =Ā (bt, bx, by, bz) and then we takeĀ cross-products of their components just like we did in three-dimensional space, so we write TijĀ = aibjĀ āˆ’Ā ajbi. Now, it’s easy to see that this combination implies that TijĀ = āˆ’ TjiĀ and that TiiĀ = 0, which is why we only have sixĀ independent numbers out of the 16 possible combinations, and which is why we’ll get a so-called anti-symmetric matrix when we organize them in a matrix. In three dimensions, the very same definition of the cross-product TijĀ gives us 9 combinations, and only 3 independent numbers, which is why we represented our ‘tensor’ as a vector too! In four-dimensional space we can’t do that: six things cannot be represented by a four-vector, so weĀ need to use thisĀ matrix, which is referred to as a tensor of the second rank in four dimensions. [When you start using words like that, you’ve come a long way, really. :-)]

[…]Ā OK. Back to our four-force. It’s easy to get a similar one-liner forĀ fyĀ and fzĀ too, of course, as well as for ft. But… Yes, ft… Is it the same thing really? Let me quickly copy Feynman’s calculation for ft:

ft

ItĀ does: remember that vƗB and v are orthogonal, and so their dot product is zero indeed. So, to make a long story short, the four equations – one for each component of the four-force vector fμ – can be summarized in the following elegant equation:

motion equation

Writing this all requires a few conventions, however. For example,Ā Fμν is a 4Ɨ4 matrix and so uν has to be written as a 1Ɨ4 vector. And the formula for theĀ fxĀ and ftĀ component also make it clear that we also want to use the +āˆ’āˆ’āˆ’ signature here, so the convention for the signs in the uνFμν product is the same as that for the scalar product aμbμ. So, in short, you really need to interpret what’s being written here.

A more important question, perhaps, is: what can we do with it? Well… Feynman’s evaluation of the usefulness of this formula is rather succinct: “Although it is nice to see that the equations can be written that way, this form is not particularly useful. It’s usually more convenient to solve for particle motions by using the F = q(E + vƗB) = (1āˆ’v2)āˆ’1/2Ā·d(m0v)/dtĀ equations, and that’s what we will usually do.”

Having said that, this formula really makes good on the promise I started my previous post with: we wanted a formula, someĀ mathematical construct, that effectively presents the electromagnetic force as oneĀ force, as one physical reality. So… Well… Here it is! šŸ™‚

Well… That’s it for today. Tomorrow we’ll talk about energy and about a veryĀ mysterious concept—the electromagnetic mass. That should be fun! So I’llĀ c u tomorrow! šŸ™‚

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 20, 2020 as a result of a DMCA takedown notice from Michael A. Gottlieb, Rudolf Pfeiffer, and The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/

Relativistic transformations of fields and the electromagnetic tensor

Pre-script (dated 26 June 2020): Our ideas have evolved into a full-blown realistic (or classical) interpretation of all things quantum-mechanical. In addition, I note the dark force has amused himself by removing some material. So no use to read this. Read my recent papers instead. šŸ™‚

Original post:

We’re going to do a very interesting piece of math here. It’s going to bring a lot of things together. The key idea is to present aĀ mathematical construct that effectively presents the electromagnetic force as oneĀ force, as one physical reality. Indeed, we’ve been saying repeatedly that electromagnetism is oneĀ phenomenon only but we’ve been writingĀ it always as something involvingĀ twoĀ vectors: he electric field vector E and the magnetic field vector B. Of course, Lorentz’ force law F = q(E + vƗB) makes it clear we’re talking one force only but… Well… There is a way of writing it all up that is much more elegant.

I have to warn you though: this post doesn’t add anything to theĀ physicsĀ we’ve seen so far: it’s all math, really and, to a large extent, math only. So if you read this blog because you’re interested in the physics only, then you may just as well skip this post. Having said that, theĀ mathematical concept we’re going to present is that of the tensorĀ and… Well… You’ll have to get to know that animal sooner or later anyway, so you may just as well give it a try right now, and see whatever you can get out of this post.

The concept of a tensor further builds on the concept of the vector, which we liked so much because it allows us to write the laws of physics as vector equations, which do notĀ change whenĀ going from one reference frame to another. In fact, we’ll see that a tensor can be described as a ‘special’ vectorĀ cross productĀ (to be precise, we’ll show that a tensor is a ‘more general’ cross product, really). So the tensor and vector concepts areĀ veryĀ closely related, but then… Well… If you think about it, the concept of a vector and the concept of a scalar are closely related, too! So we’re just moving up the value chain, so to speak: from scalar fields to vector fields to… Well… Tensor fields! And in quantum mechanics, we’ll introduce spinors, and so we also have spinor fields!Ā Having said that, don’t worry about tensor fields. Let’s first try to understand tensorsĀ tout court.Ā šŸ™‚

So… Well… Here we go.Ā Let me start with it all by reminding you of the concept of a vector, and why we like to use vectors and vector equations.

The invariance of physics and the use of vector equations

What’s a vector? You may think, naively, that any one-dimensional array of numbers is a vector. But… Well… No! In math, we may, effectively, refer to anyĀ one-dimensionalĀ array of numbersĀ as a ‘vector’, perhaps, but in physics, a vector does represent something real, something physical, and so a vector is only a vector if it transforms like a vectorĀ underĀ theĀ transformationĀ rules that apply when going from one another frame of reference, i.e. one coordinate system, to another. Examples of vectors in three dimensions are: the velocity vector v, or the momentum vector p = mĀ·v, or the position vector r.

Needless to say, the same can be said of scalars: mathematicians may define a scalar as just any real number, but it’sĀ notĀ in physics. A scalar in physics refers to something real, i.e.Ā a scalar field, like the temperature (T) inside of a block of material. In fact, think about your first vector equation: it may have been the one determining the heat flow (h), i.e. h =Ā āˆ’ĪŗĀ·āˆ‡T = (āˆ’ĪŗĀ·āˆ‚T/āˆ‚x, āˆ’ĪŗĀ·āˆ‚T/āˆ‚y, āˆ’ĪŗĀ·āˆ‚T/āˆ‚z). It immediately shows how scalar and vector fields are intimately related.

Now, when discussing the relativistic framework of physics, we introduced vectors inĀ fourĀ dimensions, i.e.Ā four-vectors.Ā The most basic four-vector is the spacetime four-vector R = (ct, x, y, z), which is often referred to as an event, but it’s just aĀ point in spacetime, really. So it’s a ‘point’ with a time as well as a spatial dimension, so it also has t in it, besides x, y and z. It is also known as the position four-vectorĀ but, again, you should think of a ‘position’ that includes time! Of course, we can re-write R as R = (ct, r), with r = (x, y, z), so here we sort of ‘break up’ the four-vector in a scalar and a three-dimensional vector, which is something we’ll do from time to time, indeed. šŸ™‚

We also have aĀ displacement four-vector, which we can write asĀ Ī”R = (cĀ·Ī”t, Ī”r). There are other four-vectors as well, including theĀ four-velocity, theĀ four-momentum and theĀ four-forceĀ four-vectors, which we’ll discuss later (in the last section of this post).

So it’s just like using three-dimensional vectors in three-dimensional physics, or ‘Newtonian’ physics, I should say: the use of four-vectors is going toĀ allow us to write the laws of physics usingĀ vector equations, butĀ in four dimensions, rather than three, so we get the ‘Einsteinian’ physics, the realĀ physics, so to speak—or the relativistically correct physics, I should say. And so these four-dimensional vector equations will also notĀ change whenĀ going from one reference frame to another, and so our four-vector will be vectors indeed, i.e. they willĀ transformĀ like a vectorĀ underĀ theĀ transformationĀ rules that apply when going from one another frame of reference, i.e. one coordinate system, to another.

What transformation? Well… In Newtonian or Galilean physics, we had translations and rotations and what have you, but what weĀ are interested in right now areĀ ‘Einsteinian’ transformations of coordinate systems, so these have to ensure that allĀ of the laws of physics that we know of, including the principle of relativity,Ā still look the same.Ā You’ve seen these transformation rules. We don’t call them the ‘Einsteinian’ transformation rules, but the LorentzĀ transformation rules, because it was a Dutch physicist (Hendrik Lorentz) who first wrote them down. So these rules are veryĀ differentĀ from the Newtonian or Galilean transformation rules which everyone assumed to be valid until the Michelson-Morley experiment unequivocally established that the speed of lightĀ did notĀ respect the Galilean transformation rules. VeryĀ different? Well… Yes. In their mathematical structure, that is. Of course, when velocities are low, i.e.Ā non-relativistic, then they yield the same result,Ā approximately, that is. However,Ā I explained that in my post on special relativity, and so I won’t dwell on that here.

Let me just jot down both sets of rules assumingĀ that the two reference frames move with respect to each other along the x- axis only, so the y- and z-component of u is zero.

Capture

The Galilean or Newtonian rules are the simple rules on the right. Going from one reference frame to another (let’s call them S and S’ respectively) is just a matter of adding or subtracting speeds: if my car goes 100 km/h, and yours goes 120 km/h, then youĀ will see my car falling behind at a speed of (minus) 20 km/h. That’s it. We could alsoĀ rotateĀ our reference frame, and our NewtonianĀ vector equationsĀ would still look the same. As Feynman notes, smilingly, it’s what a lot of armchair philosophers think relativity theory is all about, but so it’s got nothing to do with it. It’s plain wrong!

In any case, back to vectors and transformations.Ā The key to the so-calledĀ invarianceĀ of the laws of physics is the use of vectors and vector operators that transform like vectors.Ā For example, if we defined A and B as (Ax, Ay, Az) and (Bx, By, Bz), then we knew that the so-called inner productĀ A•BĀ would look the same in all rotated coordinate systems, so we can write: A•B =Ā A’•B’. So we know that if we have a product like that on both sides of an equation, we’re fine: the equation will have the same formĀ in all rotated coordinate systems. Also, the gradient, i.e. ourĀ vector operatorĀ āˆ‡Ā = (āˆ‚/āˆ‚x, āˆ‚/āˆ‚y, āˆ‚/āˆ‚z), when applied to a scalar function, gave three quantities that also transform like a vector under rotation. We also defined a vectorĀ crossĀ product, which yielded aĀ vector (as opposed to the inner product, i.e. the vectorĀ dotĀ product, which yields a scalar):

cross product

So how does this thing behave under a Galilean transformation? Well… You may or may not remember that we used this cross-product to define theĀ angular momentum L, which was a cross product of the radius vector r and the momentum vector p = mv, as illustrated below. The animation also gives the torque Ļ„, which is, loosely speaking, a measure of the turning force: it’s the cross product of r and F, i.e. the force on the lever-arm.

Torque_animation

The components of L are:

momentum angular

Now, we find that these three numbers, orĀ objectsĀ if you want, transform inĀ exactly the same wayĀ as the components of a vector.Ā However, as Feynman points out, that’s a matter of ‘luck’ really. It’s something ‘special’. Indeed,Ā you may or may not remember that we distinguished axialĀ vectors fromĀ polar vectors. L is an axial vector, while r and p are polar vectors, and so we find that, in three dimensions, the cross product of two polar vectors will always yields an axial vector. Axial vectors are sometimes referred to asĀ pseudovectors, which suggests that they are ‘not so real’ as… Well… Polar vectors, which are sometimes referred to as ‘true’ vectors. However, it doesn’t matter when doing these Newtonian or Galilean transformations: pseudo or true, both vectors transform like vectors. šŸ™‚

But so… Well… We’re actually getting a bit of a heads-up here: if we’d be mixing (or ‘crossing’) polar and axial vectors, or mixing axial vectors only, so if we’d define something involvingĀ LĀ andĀ pĀ (rather than r and p), or something involvingĀ LĀ andĀ Ļ„, then we may notĀ be so lucky, and then we’d have to carefully examine our cross-product, or whatever other product we’d want to define, because its components mayĀ notĀ behave like a vector.

Huh? Whatever other product we’d want to define? Why are you saying that?Ā Well…Ā We actuallyĀ can think of other products. For example, if we haveĀ two vectors a = (ax, ay,Ā az) and b = (bx, by, bz), then we’ll haveĀ nine possible combinations of their components, which we can write as TijĀ = aibj. So that’s like Lxy, LyzĀ and LzxĀ really. Now, you’ll say: “No. It isn’t. We don’t have nine combinations here. Just three numbers.” Well… Think about it: we actually doĀ haveĀ nineĀ LijĀ combinations too here, as we can write: LijĀ = riĀ·pj – rjĀ·pi. It just happens that, with this definition, only threeĀ of these combinations LijĀ are independent. That’s because the other six numbers are either zero or the opposite. Indeed, it’s easy to verify thatĀ LijĀ = –LjiĀ , and LiiĀ  = 0. So… Well… It turns out that the three components of our LĀ = rƗp ‘vector’ are actually a subset of a set of nineĀ LijĀ numbers.Ā So… Well… Think about it. We cannot just do whatever we want with our ‘vectors’. We need to watch out.

In fact, I do not want to get too much ahead of myself, but I can already tell you that the matrix with these nineĀ TijĀ = aibjĀ combinations is what is referred to as the tensor. To be precise, it’s referred to as a tensor of the second rank in three dimensions. The ‘second rank’, aka as ‘degree’ or ‘order’ refers to the fact that we’ve got two indices, and the ‘three dimensions’ is because we’re using three-dimensional vectors. We’ll soon see that the electromagnetic tensor is also of the second rank, but it’s a tensor in four dimensions. In any case, I should not get ahead of myself. Just note what I am saying here: the tensor is like a ‘new’ product of two vectors, a new type of ‘cross’ product really (because we’re mixing the components, so to say), but it doesn’t yield a vector: it yields aĀ matrix. For three-dimensional vectors, we get a 3Ɨ3 matrix. For four-vectors, we’ll get a 4Ɨ4 matrix. And so the full truth about our angular momentum vector L, is the following:

  1. There is a thing which we call the angular momentum tensor. It’s a 3Ɨ3 matrix, so it has nine elements which are defined as: LijĀ = riĀ·pj – rjĀ·pi. BecauseĀ of this definition, it’s an antisymmetric tensor of the second order in three dimensions, so it’s got only three independentĀ components.
  2. The three independent elements are the components of our ‘vector’ L, and picking them out and calling these three components a ‘vector’ is actually a ‘trick’ that only works in three dimensions. They really just happen toĀ transform like a vector under rotation or under whatever Galilean transformation! [By the way, do you know understand why I was saying that we can look at a tensor as a ā€˜more general’ cross product?]
  3. In fact, in four dimensions, we’ll use a similar definition and define 16 elementsĀ FijĀ as FijĀ = āˆ‡iAjĀ āˆ’Ā āˆ‡jAi, using the two four-vectors āˆ‡Ī¼Ā and Aμ (so we have 4Ɨ4 = 16 combinations indeed), out of whichĀ only sixĀ will be independent for the very same reason: we have an antisymmetric vector combination here, FijĀ = āˆ’FjiĀ and FiiĀ = 0. šŸ™‚ However, because we cannotĀ represent six independent things by four things, we doĀ notĀ get some other four-vector, and so that’s why we cannot apply the same ‘trick’ in four dimensions.

However, here I amĀ getting way ahead of myself and so… Well… Yes. Back to the main story line. šŸ™‚ So let’s try to move to the next level of understanding, which is… Well…

Because of guys like Maxwell and Einstein, we now knowĀ that rotations are part of the Newtonian world, in which time and space are neatly separated, and that things are notĀ so simple in Einstein’s world, which is the real world, as far as we know, at least! Under a Lorentz transformation, the new ā€˜primed’ space and time coordinates are a mixture of the ā€˜unprimed’ ones. Indeed, the new x’Ā is a mixture of x and t, and the new t’Ā is a mixture of x and t as well. [Yes, please scroll all the way up and have a look at the transformation on the left-hand side!]

So youĀ don’t have that under a Galilean transformation: in the Newtonian world, space and time are neatly separated, and time is absolute, i.e. it is the same regardless of the reference frame. In Einstein’s world – our world – that’s not the case: time is relative, orĀ localĀ as Hendrik Lorentz termed it quite appropriately,Ā and so it’s space-time – i.e. ā€˜some kind of union of space and time’ as Minkowski termed it – that transforms.

So that’s why physicists useĀ four-vectorsĀ toĀ keep track of things. These four-vectors always have three space-like components, but they also include one so-calledĀ time-like component.Ā It’s the only way to ensure thatĀ the laws of physics are unchanged when moving with uniform velocity.Ā Indeed, any true law of physics we write down must be arranged so that the invariance of physics (as a “fact of Nature”, as Feynman puts it) is built in, and so that’s why we use Lorentz transformations and four-vectors.

In the mentioned post, I gave a few examples illustrating how the Lorentz rules work. Suppose we’re looking at some spaceship that is moving at half the speed of light (i.e. 0.5c) and that, inside the spaceship, some object is also moving at half the speed of light, as measured in the reference frame of the spaceship, then we get the rather remarkable result that, from ourĀ point of view (i.e. ourĀ reference frame as observer on the ground), that object is notĀ going as fast as light, as Newton or Galileo – and most present-day armchair philosophers šŸ™‚ – would predict (0.5cĀ +Ā 0.5cĀ = c). We’d see it move at a speed equal to vĀ =Ā 0.8c. Huh?Ā How do we know that? Well… We can derive a velocity formula from the Lorentz rules:

Capture

So now you can just put in the numbers now:Ā vxĀ = (0.5c + 0.5c)/(1 + 0.5Ā·0.5) = 0.8c. See?

Let’s do another example. Suppose we’re looking at a light beam inside the spaceship, so something that’s traveling at speed c itself in the spaceship. How does that look to us? The Galilean transformation rules say its speed should be 1.5c, but that can’t be true of course, and the Lorentz rules save us once more: vxĀ = (0.5cĀ + c)/(1 + 0.5Ā·1) = c, so it turns out that the speed of light doesĀ notĀ depend on the reference frame: it looks the same – both to the man in the ship as well as to the man on the ground. As Feynman puts it: “This is good, for it is, in fact, what the Einstein theory of relativity was designed to do in the first place—so it hadĀ betterĀ work!” šŸ™‚

So let’s now apply relativity to electromagnetism. Indeed, that’s what this post is all about! However, before I do so, let me re-write the Lorentz transformation rules forĀ cĀ = 1. We can equate the speed of light to one, indeed, when measure time and distance in equivalent units. It’s just a matter of ditching our seconds for meters (so our time unit becomes the time that light needs to travel a distance of one meter), or ditching our meters for seconds (so our distance unit becomes the distance that light travels in one second). You should be familiar with this procedure. If not, well… Check out my posts on relativity. So here’s the same set of rules for cĀ = 1:

Lorentz rules

They’re much easier to remember and work with, and so that’s good, because now we need toĀ look at how these rules work with four-vectors and the various operations and operators we’ll be defining on them. Let’s look at that step by step.

Electrodynamics in relativistic notation

Let me copy the UniversalĀ Set of Equations and Their Solution once more:

frame

The solution for Maxwell’s equations is given in terms of the (electric) potential Φ and the (magnetic) vectorĀ potential A. I explained that in my post on this, so I won’t repeat myself too much here either. The only point you should note is that this solution is the result of a special choice of Φ andĀ A, which we referred to as the Lorentz gauge.Ā We’ll touch upon this condition once more, so just make a mental note of it.

Now, E and B do not correspond to four-vectors: theyĀ dependĀ on x, y, z and t, but they haveĀ threeĀ components only:Ā Ex, Ey,Ā Ez, and Bx, By, and BzĀ respectively. So we have sixĀ independent terms here, rather thanĀ four things that, somehow, we could combine into some four-vector. [Does this ring a bell? It should. :-)] Having said that, it turns out that weĀ canĀ combine Φ andĀ AĀ into a four-vector, which we’ll refer to as the four-potentialĀ and which we’llĀ will write as:

Aμ = (Φ, A) = (Φ, Ax, Ay, Az) = (At, Ax, Ay, Az) with At = Φ.

So that’s a four-vector just likeĀ R = (ct, x, y, z).

How do we know that Aμ is a four-vector? Well… Here I need to say a few things about those Lorentz transformation rules and, more importantly, about the required condition ofĀ invarianceĀ under a Lorentz transformation. So, yes, here we need to dive into the math.

Four-vectors and invariance under Lorentz transformations

When you were in high-school, you learned how toĀ rotateĀ your coordinate frame. You also learned that the distance of a point from the origin does not change under a rotation, so you’d writeĀ r’2Ā = x’2Ā + y’2Ā + z’2Ā =Ā r2Ā =Ā x2Ā + y2Ā + z2, and you’d say that r2Ā is an invariant quantity under a rotation.Ā Indeed, transformations leave certain things unchanged. From the Lorentz transformation rules itself, it is easy to see that

cĀ·t’2 – x’2 – y’2 –z ‘2 = cĀ·t2 –x2 – y2Ā  – z2, or,

if cĀ = 1, thatĀ t’2 – x’2 – y’2 – z’2 = t2 – x2 – y2Ā  – z2,

is anĀ invariantĀ under a Lorentz transformation. We found the same for the so-calledĀ spacetimeĀ interval Ī”s2Ā  =Ā Ī”r2 – cĪ”t2, which we write as Ī”s2Ā  =Ā Ī”r2 – Ī”t2Ā as we chose our time or distance units such that cĀ = 1. [Note that, from now on, we’ll assume that’s the case, so cĀ = 1 everywhere. We can always change back to our old units when we’re done with the analysis.] Indeed, such invariance allowed us to define spacelike,Ā timelikeĀ and lightlikeĀ intervals using the so-called light cone emanating from a single event and traveling in all directions.

You should note that, for four-vectors, we do not have a simple sum of three terms. Indeed, we don’t write x2Ā + y2Ā + z2Ā but t2 – x2 – y2Ā  – z2. So we’ve got a +āˆ’āˆ’āˆ’ thing here or, it’s just another convention, we could also work with a āˆ’+++ sum of terms. The convention is referred to as the signature, and we will use the so-calledĀ metric signature here, which is +āˆ’āˆ’āˆ’. Let’s continue the story.Ā Now, all four-vectors aμ = (at, ax, ay, az) have this property that:

at2 – ax2 – ay2 – az2 = at2 – ax2 – ay2Ā  – az2.

[The primed quantities are, obviously, the quantities as measured in the other reference frame.]Ā So. Well… Yes. šŸ™‚ But… Well… Hmm… We can sayĀ that our four-potential vector is a four-vector, butĀ so we still have toĀ proveĀ that. So we need to prove that Φ’2 – Ax2 – Ay2 – Az2 = Φ2 – Ax2 – Ay2Ā  – Az2Ā for ourĀ four-potential vectorĀ Aμ = (Φ, A). So… Yes… How can we do that? The proof isĀ notĀ so easy, but you need to go through it as it will introduce some more concepts and ideas you need to understand.

In my post on the Lorentz gauge, I mentioned that Maxwell’s equations can be re-written in terms of Φ andĀ A, rather than in terms of E and B. The equations are:

Equations 2

The expression look rather formidable, but don’t panic: just look at it. Of course, you need to be familiar with the operators that are being used here, so that’s the Laplacian āˆ‡2Ā and the divergence operator āˆ‡ā€¢ that’s being applied to the scalar Φ and the vector A. I can’t re-explain this. I am sorry. Just check my posts on vector analysis.Ā You should also look at the third equation: that’s just the Lorentz gauge condition, which we introduced when derivingĀ these equations from Maxwell’s equations. Having said that, it’s the first and second equation which describe Φ and A as a function of the charges and currents in space, and so that’s what matters here. So let’s unfold the first equation. It says the following:

potential formula

In fact, if we’d be talkingĀ freeĀ or empty space, i.e. regions where there are no charges and currents, then the right-hand side would be zero and this equation would then represent a wave equation, so some potential Φ that is changing in time and moving out at the speed c. Here again, I am sorry I can’t write about this here: you’ll need to check one of my posts on wave equations. If you don’t want to do that, you should believe me when I say that, if you see an equation like this:

f8then the functionĀ ĪØ(x, t) must be some function

solution

Now, that’s a function representing a wave traveling at speed c, i.e. the phase velocity. Always? Yes.Ā Always! It’s got to do with the x āˆ’ ct and/or x +Ā ctĀ  argument in the function. But, sorry, I need to move on here.

The unfolding of the equation with Φ makes it clear that we have four equations really. Indeed, the second equation is three equations: one for Ax, one forĀ Ay, and one forĀ AzĀ respectively. The four quantities on the right-hand side of these equations are ρ, jx, jyĀ and jzĀ respectively, divided by ε0, which is a universal constant which does notĀ change when going from one coordinate system to another.Ā Now, the quantities ρ, jx, jyĀ and jzĀ transform like a four-vector. How do we know that? It’s just the charge conservation law. We used it when solving the problem of the fields around a moving wire, when we demonstrated theĀ relativityĀ of the electric and magnetic field. Indeed, the relevant equations were:

Lorentz j and rho

You can check that against the Lorentz transformation rules forĀ cĀ = 1. They’re exactly the same, but so we chose t = 0, so the rules are even simpler. Hence, the (ρ,Ā jx, jy, jz) vector is, effectively, aĀ four-vector, and we’ll denote it by jμ = (ρ, j).Ā I nowĀ need to explain something else. [And, yes,Ā I know this is becoming a veryĀ long story but… Well… That’s how it is.]

It’s about our operatorsĀ āˆ‡, āˆ‡ā€¢, āˆ‡Ć— and āˆ‡2Ā , so that’s the gradient, theĀ divergence, curlĀ and LaplacianĀ operator respectively: they all have a four-dimensional equivalent. Of course, that won’t surprise you. 😦 Let me just jot all of them down, so we’re done with that, and then I’ll focus on the four-dimensional equivalent of the LaplacianĀ Ā āˆ‡ā€¢āˆ‡ =Ā āˆ‡2Ā , which is referred to as theĀ D’Alembertian, and which is denoted byĀ ā–”2, because that’s the one we need to prove that our four-potential vector is a realĀ four-vector. [I know: ā–”2Ā is a tiny symbol for a pretty monstrous thing, but I can’t help it: my editor tool is pretty limited.]

Four-vectors

Now, we’re almost there. Just hang in for a little longer. It should be obvious that we can re-write those two equations with Φ, A, ρ and j, as:

Formula d'alembertian 2

Just to make sure, let me remind you that Aμ = (Φ, A) and thatĀ jμ = (ρ, j). Now, our new D’Alembertian operator is just an operator—a prettyĀ formidableĀ operator but, still, it’s an operator, and so itĀ doesn’t change when the coordinate system changes, so the conclusion is that,Ā IFĀ jμ = (ρ, j) is a four-vector – which it is – and, therefore, transforms like a four-vector,Ā THENĀ the quantities Φ, Ax, Ay, and AzĀ must also transformĀ like a four-vector, which means they areĀ (the components of) a four-vector.

So… Well… Think about it, but not too long, because it’s just an intermediate result we had to prove. So that’s done. But we’re not done here. It’s just the beginning, actually. :-/ Let me repeat our intermediate result:

Aμ = (Φ, A) is a four-vector. We call it the four-potential vector.

OK. Let’s continue. Let me firstĀ draw your attention to that expression with the D’Alembertian above. Which expression? This one:

Formula d'alembertian 2

What about it? Well… You should note thatĀ the physics of that equation is just the same as Maxwell’s equations. So it’s one equation only, but it’s got it all.

It’s quite a pleasure to re-write it in such elegant form. Why? Think about it: it’s a four-vector equation: we’ve got a four-vector on the left-hand side, and a four-vector on the right-hand side. Therefore, this equation is invariant under a transformation. So, therefore,Ā it directly shows the invariance of electrodynamics under the Lorentz transformation.

Huh? Yes. You may think about this a little longer. šŸ™‚

To wrap this up, I should also note that we can also express the gauge condition using our new four-vector notation. Indeed, we can write it as:

Lorentz condition

It’s referred to as the Lorentz condition and it is, effectively, a condition for invariance, i.e. it ensures that the four-vector equation above does stay in the form it is in for all reference frames. Note that we’re re-writing it using the four-dimensional equivalent of the divergence operatorĀ āˆ‡ā€¢, but so we don’t have a dot between āˆ‡Ī¼Ā and Aμ. In fact, the notation is pretty confusing, and it’s easy to think we’re talking some gradient, rather than the divergence. So let me therefore highlight the meaning of both once again. It looks the same, but it’s two veryĀ different things: the gradient operates on a scalar, while the divergence operates on a (four-)vector. Also note the +āˆ’āˆ’āˆ’ signature is only there for theĀ gradient, not for the divergence!

example

You’ll wonder why they didn’t use some • orĀ āˆ— symbol, and the answer: I don’t know. I know it’s hard to keep inventing symbols for all these different ‘products’ – the āŠ— symbol, for example, is reserved forĀ tensorĀ products, which we won’t get into – but… Well… I think they could have done something here. 😦

In any case… Let’s move on. Before we do, please note that we can also re-write our conservation law for electric charge using our new four-vector notation. Indeed, you’ll remember that we wrote that conservation law as:

conservation law

Using our new four-vector operator āˆ‡Ī¼, we can re-write that as āˆ‡Ī¼jμ = 0. So all of electrodynamics can be summarized in the two equations only—Maxwell’s law and the charge conservation law:

all

OK. We’re now ready to discuss the electromagnetic tensor. [I know… This is becoming an incredibly long and incredibly complicated piece but, ifĀ you get through it, you’ll admitĀ it’s really worth it.]

The electromagnetic tensor

The whole analysis above was done in terms of the Φ and A potentials. It’s time to get back to our field vectorsĀ E and B. We know we can easily get them from Φ and A, using the rules we mentioned as solutions:

E and B solutions

These two equations shouldĀ notĀ look as yet another formula. They are essential, and you should be able to jot them down anytime anywhere. They should be on your kitchen door, in your toilet and above your bed. šŸ™‚Ā For example, the second equation gives us theĀ components of the magnetic field vector B:

B field components

Now, look at these equations. The x-component is equal to a couple of terms that involve only y– and z-components. The y-component is equal to something involving only xĀ and z.Ā Finally, the z-component only involves x and y. Interesting. Let’s define a ‘thing’ we’ll denote by FzyĀ and define as:

F definition

So now we can write: BxĀ = Fzy,Ā ByĀ = Fxz, andĀ BzĀ = Fxy. Now look at our equation for E. It turns out the components of E are equal to things likeĀ Fxt, FytĀ and Fzt! Indeed,Ā FxtĀ =Ā āˆ‚Ax/āˆ‚tĀ āˆ’Ā āˆ‚At/āˆ‚x = Ex!

But… Well… No. 😦 The sign is wrong!Ā ExĀ = āˆ’āˆ‚Ax/āˆ‚tāˆ’āˆ‚At/āˆ‚x, so we need to modify our definition of Fxt. When the t-component is involved, we’ll define our ‘F-things’ as:

time f

So we’ve got a plus instead of a minus. It looks quite arbitrary but, frankly, you’ll have to admit it’s sort of consistent with our +āˆ’āˆ’āˆ’ signatureĀ for our four-vectors and, in just a minute, you’ll see it’s fully consistent with our definition of the four-dimensional vector operatorĀ āˆ‡Ī¼Ā = (āˆ‚/āˆ‚t, āˆ’āˆ‚/āˆ‚x,Ā āˆ’āˆ‚/āˆ‚y, āˆ’āˆ‚/āˆ‚z). So… Well… Let’s go along with it.

What about the Fxx, Fyy, FzzĀ and FttĀ terms? Well…Ā FxxĀ =Ā āˆ‚Ax/āˆ‚x āˆ’Ā āˆ‚Ax/āˆ‚x = 0, and it’s easy to see that FyyĀ and FzzĀ are zero too. ButĀ Ftt? Well… It’s a bit tricky but, applying our definitions carefully, we see that FttĀ must be zero too. In any case, theĀ FttĀ = 0 will become obvious as we will be arranging these ‘F-things’ in a matrix, which is what we’ll do now.Ā [Again: does this ring a bell? If not, it should. :-)]

Indeed, we’ve got sixteen possible combinations here, which Feynman denotes as Fμν, which is somewhat confusing, because Fμν usually denotes theĀ 4Ɨ4 matrixĀ representing all of these combinations. So let me use the subscripts i and j instead, and define FijĀ as:

FijĀ =Ā āˆ‡iAjĀ āˆ’Ā āˆ‡jAi

with āˆ‡iĀ being the t-, x-, y- orĀ z-component of āˆ‡Ī¼Ā =Ā (āˆ‚/āˆ‚t, āˆ’āˆ‚/āˆ‚x,Ā āˆ’āˆ‚/āˆ‚y,Ā āˆ’āˆ‚/āˆ‚z) and, likewise, AiĀ being the t-, x-, y- orĀ z-component of Aμ =Ā (Φ, Ax,Ā Ay,Ā Az). Just check it: FzyĀ = āˆ’āˆ‚Ay/āˆ‚z +Ā āˆ‚Az/āˆ‚y = āˆ‚Az/āˆ‚y āˆ’ āˆ‚Ay/āˆ‚z = Bx, for example, andĀ FxtĀ =Ā āˆ’āˆ‚Ī¦/āˆ‚x āˆ’Ā āˆ‚Ax/āˆ‚tĀ = Ex. So theĀ +āˆ’āˆ’āˆ’ conventionĀ works. [Also note that it’s easier now to see that FttĀ = āˆ‚Ī¦/āˆ‚t āˆ’ āˆ‚Ī¦/āˆ‚t = 0.]

We can now arrange the FijĀ in a matrix. This matrix is antisymmetric, because FijĀ = – Fji, and its diagonal elements are zero. [For those of you who love math: note that the diagonal elements of an antisymmetric matrix are always zero because of the FijĀ = – FjiĀ constraint: just use k = i = j in the constraint.]

Now that matrix is referred to as the electromagnetic tensor and it’s depicted below (we pluggedĀ cĀ back in, remember thatĀ B’s magnitude is 1/c times E’s magnitude).

electromagnetic tensor final

So… Well… Great ! We’re done! Well… Not quite. šŸ™‚

We can get this matrix in a number of ways. The least complicated way is, of course, just to calculate all FijĀ components and them put them in a [Fij] matrix using theĀ iĀ as the row number and theĀ jĀ as the column number. You need to watch out with the conventions though, and soĀ i and j startĀ onĀ t and end on z. šŸ™‚

The other way to do it is to write theĀ āˆ‡Ī¼Ā = (āˆ‚/āˆ‚t, āˆ’āˆ‚/āˆ‚x,Ā āˆ’āˆ‚/āˆ‚y,Ā āˆ’āˆ‚/āˆ‚z) operator as a 4Ɨ1 column vector, which you then multiply with the four-vector Aμ written as a 4Ɨ1 row vector. So āˆ‡Ī¼Aμ is then aĀ 4Ɨ4 matrix, which we combine with its transpose, i.e.Ā (āˆ‡Ī¼Aμ)T, as shown below. So what’s written below is (āˆ‡Ī¼Aμ) āˆ’ (āˆ‡Ī¼Aμ)T.

matrix

If you google, you’ll see there’s more than one way to go about it, so I’d recommend you just go through the motions and double-check the whole thing yourself—and please do let me know if you find any mistake! In fact, the Wikipedia article on the electromagnetic tensor denotes the matrix above asĀ Fμν, rather than asĀ Fμν, which is the sameĀ tensorĀ but in its so-calledĀ covariantĀ form, but so I’ll refer you to that article as I don’t want to make things even more complicated here! As said, there’s differentĀ conventionsĀ around here, and so you need to double-check what is what really. šŸ™‚

Where are we heading with all of this? The next thing is to look at theĀ LorentzĀ transformation of theseĀ FijĀ =Ā āˆ‡iAjĀ āˆ’Ā āˆ‡jAiĀ components, because then we know how our E and B fields transform. Before we do so, however, we should note the more general results and definitions which we obtained here:

1. The Fμν matrix (a matrix is just a multi-dimensional array, of course) is a so-called tensor. It’s a tensor of the second rank, because it has two indices in it. We think of it as a very special ‘product’ of two vectors, not unlike the vector cross product aĀ Ć— b, whose components were also defined by a similar combination of the components of a and b. Indeed, we wrote:

cross product

So one should think of a tensor as “another kind of cross product” or, preferably, and as Feynman puts it, as a “generalization of the cross product”.

2.Ā In this case, the four-vectors are āˆ‡Ī¼Ā =Ā (āˆ‚/āˆ‚t, āˆ’āˆ‚/āˆ‚x,Ā āˆ’āˆ‚/āˆ‚y,Ā āˆ’āˆ‚/āˆ‚z) and Aμ =Ā (Φ, Ax,Ā Ay,Ā Az). Now, you will probably say that āˆ‡Ī¼Ā is an operator, not a vector, and you are right. However, we know that āˆ‡Ī¼Ā behaves like a vector, and so this is just a special case. The point is: because the tensor is based on four-vectors, the Fμν tensor is referred to as a tensor of the second rank in four dimensions. In addition, because of the FijĀ = – FjiĀ result, Fμν is an asymmetric tensor of the second rank in four dimensions.

3.Ā Now, the whole point is to examine how tensors transform. We know that the vector dot product, aka the inner product, remains invariantĀ under a Lorentz transformation, both in three as well as in four dimensions, but what about the vector cross product, and what about the tensor? That’s what we’ll be looking at now.

The Lorentz transformation of the electric and magnetic fields

Cross products are complicated, and tensors will be complicated too. Let’s recall our example in three dimensions, i.e. the angular momentum vectorĀ L, which was a cross product of the radius vector r and the momentum vector p = mv, as illustrated below (the animation also gives the torque Ļ„, which is, loosely speaking, a measure of the turning force).

Torque_animation

The components of L are:

momentum angular

Now, this particular definition ensures that LijĀ turns out to be an antisymmetric object:

three-vector

So it’s a similar situation here. We have nineĀ possible combinations, but only threeĀ independent numbers. So it’s a bit like our tensor in four dimensions: 16 combinations, but only 6 independent numbers.

Now, it so happens that that these three numbers, orĀ objectsĀ if you want, transform inĀ exactly the same wayĀ as the components of a vector.Ā However, as Feynman points out, that’s a matter of ‘luck’ really. In fact, Feynman points out that, when we have two vectors a = (ax, ay,Ā az) and b = (bx, by, bz), we’ll haveĀ nineĀ products TijĀ = aibjĀ which will also form aĀ tensorĀ of the second rank (cf. the two indices) but which, in general, will not obey the transformation rules we got for theĀ angular momentumĀ tensor, whichĀ happenedĀ to be an antisymmetric tensor of the second rank in three dimensions.

To make a long story short, it’s not simple in general, and surely not here: with E and B, we’ve gotĀ sixĀ independent terms, and so we cannotĀ represent six things by four things, so the transformation rules for E and B will differ from those for a four-vector. So what areĀ they then?

Well… Feynman first works out the rules for the general antisymmetric vector combination GijĀ = aibjĀ āˆ’ ajbi, withĀ aiĀ and bjĀ the t-, x-, y- or z-component of the four-vectors aμ = (at, ax, ay, az) and bμ = (bt, bx, by, bz) respectively. The idea is to first get some general rules, and then replace GijĀ = aibjĀ āˆ’ ajbiĀ by FijĀ =Ā āˆ‡iAjĀ āˆ’Ā āˆ‡jAi, of course! So let’s apply the Lorentz rules, which – let me remind you – are the following ones:

Lorentz rules

So we get:

set 1

The rest is all very tedious: you just need to plug these things into the variousĀ GijĀ = aibjĀ āˆ’ ajbiĀ formulas. For example, for G’tx, we get:

G1

Hey! That’s justĀ G’tx, so we find that G’txĀ =Ā Gtx! What about the rest? Well… ThatĀ yields something different. Let me shorten the story by simply copying Feynman here:

resulsts

So… Done!

So what?

Well… Now we just substitute. In fact, thereĀ areĀ two alternative formulations of the Lorentz transformations of E and B. They are given below (note the units are such thatĀ c = 1):

result 1 result 2

In addition, there is a third equivalent formulation which is more practical, and also simpler, even if it puts theĀ c‘s back in. It re-defines the field components, distinguishing only two:

  1. The ‘parallel’ components E||Ā and B||Ā along theĀ x-direction ( because they are parallel to the relative velocity of the S and S’ reference frames), and
  2. The ‘perpendicular’ or ‘total transverse’ componentsĀ E⊄ and B⊄, which are the vector sums of the y- and z-components.

So that gives us four equations only:

result 3

And, yes, we areĀ done now. This is the Lorentz transformation of the fields. I am sure it has left you totally exhausted. Well… If not… […] It sure left me totally exhausted. šŸ™‚

To lighten things up, let me insert an image of how the transformed field E actually looks like. The first image is the reference frame of a charge itself: we have a simple Coulomb field. The second image shows the charge flying by. Its electric field is ‘squashed up’. To be precise, it’s just like the scale ofĀ xĀ is squashed up by a factorĀ ((1āˆ’v2/c2)1/2. Let me refer you to Feynman for the detail of the calculations here.

field

OK. So that’s it. You may wonder: what about that promise I made? Indeed, when I started this post, I said I’dĀ present aĀ mathematical construct that presents the electromagnetic force as oneĀ force only, as one physical reality, but so we’re back writing all of it in terms of twoĀ vectors—the electric field vector E and the magnetic field vector B. Well… What can I say? IĀ didĀ present the mathematical construct: it’s the electromagnetic tensor. So it’s that antisymmetric matrix really, which one can combine with aĀ transformation matrixĀ embodying the Lorentz transformation rules. So, IĀ didĀ what I promised to do. But you’re right: IĀ amĀ re-presenting stuff in the old style once again.

The second objection that you may have—in fact, that youĀ shouldĀ have, is that all of this has been rather tedious. And you’re right.Ā The whole thing just re-emphasizes the value of using the four-potential vector. It’s obviouslyĀ muchĀ easier to takeĀ thatĀ vector from one reference frame to another – so we just apply the Lorentz transformation rules to Aμ = (Φ, A) and get Aμ‘ = (Φ’, A’) from it – and then calculate E’ and B’ from it, rather than trying to remember those equations above.Ā However, that’s not the point, or…

Well… It is and it isn’t. We wanted to get away from thoseĀ twoĀ vectors E and B, and show thatĀ electromagnetism is reallyĀ oneĀ phenomenon only, and so that’s where the concept of the electromagnetic tensor came in. There were two objectives here: the first objective was to introduce you to the concept of tensors, which we’ll need in the future. The second objective was to show you that, while Lorentz’ force law – F = q(E + vƗB) makes it clear we’re talking one force only, there is a way of writing it all up that is much more elegant.

I’ve introduced the concept of tensors here, so the first objective should have been achieved. As for the second objective, I’ll discuss that in my next post, in which I’ll introduce the four-velocity vector μμ as well as the four-force vectorĀ fμ. It will explain the following beautiful equation of motion:

motion equation

NowĀ thatĀ looks very elegant and unified, doesn’t it? šŸ™‚

[…] Hmm… No reaction. I know… You’re tired now, and you’re thinking: yet another way of representing the same thing? Well… Yes! So…

OK… Enough for today. Let’s follow up tomorrow.

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 20, 2020 as a result of a DMCA takedown notice from Michael A. Gottlieb, Rudolf Pfeiffer, and The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 20, 2020 as a result of a DMCA takedown notice from Michael A. Gottlieb, Rudolf Pfeiffer, and The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 20, 2020 as a result of a DMCA takedown notice from Michael A. Gottlieb, Rudolf Pfeiffer, and The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/

Electric circuits (2): Kirchoff’s rules, and the energy in a circuit

My previous post was long and tedious, and all it did was presenting the three (passive) circuit elements as well as the concept of impedance. It show the inner workings of these little devices are actually quite complicated.Ā Fortunately, the conclusions were very neat and short: for all circuit elements, we have a very simple relationship between (a) the voltage acrossĀ the terminals of the element (V) and (b) the current that’s goingĀ throughĀ the circuit element (I). We found they are always in some ratio, which is referred to as the impedance, which we denoted by Z:

Z =Ā V/I ⇔ V = Iāˆ—Z

So it’s a ‘simple’ ratio, indeed. But… Well… Simple and not simple. It’s a ratio of two complex numbers and, therefore, it’s aĀ complexĀ number itself. That’s why I use the āˆ— symbol when re-writing the Z =Ā V/IĀ formula as V = Iāˆ—Z, so it’s clear we’re talking a product of two complex numbers). This ‘complexity’ is best understood by thinking of the voltage and the current asĀ phase vectorsĀ (orĀ phasorsĀ as engineers call them). Indeed, instead of using the sinusoidal functions we are used to, so that’s

  • V = V0Ā·cos(ωt + ĪøV),
  • I = I0Ā·cos(ωt + ĪøI), and
  • Z = Z0Ā·cos(ωt + Īø) = (V0/I0)Ā·cos(ωt + ĪøV āˆ’Ā ĪøI),

we preferred theĀ complex orĀ vectorĀ notation, writing:

  • VĀ =Ā |V|ei(ωt +Ā ĪøV) = V0Ā·ei(ωt +Ā ĪøV)
  • IĀ =Ā |I|ei(ωt +Ā ĪøI)Ā = I0Ā·ei(ωt +Ā ĪøI)
  • Z = |ZIei(ωt +Ā Īø)Ā = Z0Ā·ei(ωt +Ā Īø)Ā = (V0/I0)Ā·ei(ωt +Ā ĪøV āˆ’Ā ĪøI)

For the three circuit elements, we found the followingĀ solutionĀ for ZĀ in terms of the previously definedĀ propertiesĀ of the respective circuit elements, i.e. their resistanceĀ (R),Ā capacitanceĀ (C), and inductanceĀ (L) respectively:

  1. For a resistor, we have Z(resistor) = ZRĀ = R
  2. For an capacitor, we have Z(capacitor) = ZCĀ = 1/iωC = –i/(ωC)
  3. For an inductor, we haveĀ Z(inductance) = ZL= iωL

We also explained what these formulas meant, using graphs like the ones below:

  1. The graph on the left-hand side gives you the ratio of theĀ peakĀ voltage andĀ peakĀ current for the three devices as a function of C, L, R and ω respectively.
  2. The graph on the right-hand side shows you the relationship between theĀ phaseĀ of the voltage and the current for a capacitor and for an inductor. [For a resistor, the phases are the same, so no need for a graph. Also note that the lag of theĀ phaseĀ of the current vis-Ć”-vis the voltage phase is 90 degrees for an inductor, while it’s 270 degrees for a capacitorĀ (which amounts to the current leading the voltage with a 90° phase difference).]

Capture800px-VI_phase

The inner workings of our circuit elements are all wonderful and mysterious, and so we spent a lot of time writing about them. That’s finished now. The summary about describes all of them in very simple terms, relating the voltage and currentĀ phasorĀ through the concept of impedance, which is just a ratio—albeit aĀ complexĀ ratio.

As the graphs above suggest, we can build all kinds of crazy circuits now, and the idea of resonanceĀ as we’ve learned it when studying the behavior of waves will be particularly relevant when discussing circuits that are designed to filter certain frequenciesĀ or, the opposite, to amplifyĀ some. We won’t go that far in this post, however, as I just want to explain the basic rules one needs to know when looking at a circuit, i.e.Ā Kirchoff’s circuit laws.Ā There are two of them:

1. Kirchoff’s Voltage Law (KCL): The sum of the voltage drops around any closed path is zero.

The principle is illustrated below. It doesn’t matter whether or not we have other circuits feeding into this one: Kirchoff’s Voltage LawĀ (KCL) remains valid.

KVR

We can write this law using the concept of circulation once again or, what you’ll probably like more, just using plain summation:Ā integral

KVL

2. Kirchoff’s Current Law (KCL): The sum of the currents into any node is zero.

This law is written and illustrated as follows:

KCL

KCL illus

KCR

This Law requires some definition of aĀ node, of course. Feynman defines a node as any set of terminals such as a, b, c, d in the illustration above which are connected. So it’s a set of connected terminals.

Now, I’ll refer you Feynman for some practicalĀ examples. The circuit below is one of them. It looks complicated but it all boils down to solving a set of linear equations. So… Well… That’s it, really. We’re done! We should do the exercises, of course, but then we’re too lazy for that, I guess. šŸ™‚ So we’re done!

example

Well… Almost. I also need to mention how one can reduceĀ complicated circuits by combining parallel impedances, using the following formula:

parallel Z

And then another powerful idea is the idea ofĀ equivalentĀ circuits. The rules for this are as follows:

  1. Any two-terminal network of passive elements is equivalent to (and, hence, can be replaced by) anĀ effectiveĀ impedance (Zeff).
  2. Any two-terminal network of passive elements is equivalent to (and, hence, can be replaced by) aĀ generator in series with an impedance.

These two principles are illustrated below: (a) is equivalent to (b) in each diagram.

P2P1

The related formulas are:

  1. I = Ɛ/Zeff
  2. VnĀ = ƐeffĀ āˆ’ Ināˆ—Zeff

Last but not least, I need to say something about the energyĀ in circuits. As we noted in our previous post, the impedance will consist of a real and an imaginary part. We write:

Z = R + iĀ·X

This gives rise to the following powerful equivalence:Ā any impedance is equivalent to a series combination of a pure resistance and a pure reactance, as illustrated below (the ≔ sign stands for equivalence):

energy

Of course, because this post risks becoming too short šŸ™‚ I need to insert some more formulas now. IfĀ Z = R + iĀ·X is the impedance of the whole circuit, then the whole circuit can be summarized in the following equation:

Ɛ = Iāˆ—ZĀ = Iāˆ—(R + iĀ·X)

Now, if we bring the analysis back to the realĀ parts of this equation, then we may write our current as I = I0Ā·cos(ωt). This implies we chose a t = 0 point soĀ ĪøIĀ = 0. [Note that this is somewhat different than what we usually do: we usually chose our t = 0 point such thatĀ ĪøVĀ = 0, but it doesn’t matter.] The real emf is then going to be the real part of Ɛ = Iāˆ—ZĀ = Iāˆ—(R + iĀ·X), so we’ll write it as ʐ (no bold-face), and it’s going to be the real part of that expression above, which we can also write as:

Ɛ = Iāˆ—ZĀ = I0Ā·ei(ωt)Ā āˆ—(R + iĀ·X)

So ʐ is the real part of this Ɛ and, you should check, it’s going to be equal to:

ʐ =Ā I0Ā·RĀ·cos(ωt) āˆ’ I0Ā·XĀ·sin(ωt)

The two terms in this equation represent the voltage drops across the resistance R and the reactance X in that illustration above. […] Now that I think of it, in line with the -or and -ance convention for circuit elements and their properties, should we, perhaps, sayĀ resistorĀ andĀ reactor in this case? šŸ™‚ […] OK. That’s a bad joke. [I don’t seem to have good ones, isn’t it?] šŸ™‚

Jokes aside, we see thatĀ the voltage drop across the resistance is in phase with the current (because it’s a simple cosine function of ωt as well), while the voltage drop across the purely reactive part is out of phase with the current (as you know, the sine and cosine are the same function, but with a phase difference of π/2 indeed).

You’ll wonder where are we going with this, so let me wrap it all up. You know the power is the emf times the current, and so let’s integrate this thing over one cycle to get the average rate (and then I mean aĀ time rate of change) of the energy that getsĀ lostĀ in the circuit. So we need to solve the following integral:

integral 2

This may look like a monster, but if you look back at your notes from your math classes, you should be able to figure it out:

  1. The first integral is (1/2)I02Ā·R..
  2. The second integral is zero.

So what? Well… Look at it! It means that the (average) energy loss in a circuit with impedance Z = R + iĀ·X onlyĀ depends on the real part of Z, which is equal to I02Ā·R/2. That’s, of course, how we want it to be: ideal inductances and capacitorsĀ storeĀ energy when being powered, and give whatever they stored when ‘charging’ back to the circuit when the current reverses direction.

So it’s a nice result, because it’s consistent with everything. Hmm… Let’s double-check though… Is it also consistent with the power equation for a resistor which, remember, is written as: P = VĀ·I = IĀ·RĀ·I = I2Ā·R. […] Well… What about the 1/2 factor?

Well… Think about it. I isĀ a sine or a cosine here, and so we want the average value of its square, so that’s 〈cos2(ωt)〉 = 1/2.

Done!Ā šŸ™‚

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 17, 2020 as a result of a DMCA takedown notice from Michael A. Gottlieb, Rudolf Pfeiffer, and The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 17, 2020 as a result of a DMCA takedown notice from Michael A. Gottlieb, Rudolf Pfeiffer, and The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 20, 2020 as a result of a DMCA takedown notice from Michael A. Gottlieb, Rudolf Pfeiffer, and The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/

Electric circuits (1): the circuit elements

OK. No escape. It’s part of physics. I am not going to go into the nitty-gritty of it all (because this is a blog about physics, not about engineering)Ā but it’s good to review the basics, which are, essentially, Kirchoff’s rules. Just for the record, GustavĀ Kirchhoff was a German genius who formulated these circuit laws while he was still a student, when he was like 20 years old or so. He did it as aĀ seminar exerciseĀ 170 years ago, and then turned it into doctoral dissertation. Makes me think of that Dire Straits song—That’s the way you do it—Them guys ain’t dumb. šŸ™‚

So this post is, in essence, just an ‘explanation’ of Feynman’s presentation of Kirchoff’s rules, so I am writing this post basically for myself, so as to ensure I am not missing anything. To be frank, Feynman’s use of notation when working with complex numbers is confusing at times and so, yes, I’ll do some ‘re-writing’ here.Ā The nice thing about Feynman’s presentation of electrical circuits is that he sticks to Maxwell’s Laws when describing all ideal circuit elements, so he keeps using line integrals of the electric field EĀ around closed paths (that’s what a circuit is, indeed) to describe the so-calledĀ passive circuit elements, and he also recapitulates the idea of theĀ electromotive forceĀ when discussing the so-calledĀ activeĀ circuit element, so that’s the generator. That’s nice, because it links it all with what we’ve learned so far, i.e. theĀ fundamentalsĀ as expressed in Maxwell’s set of equations. Having said that, I won’t make that link here in this post, because I feel it makes the whole approach rather heavy.

OK. Let’s go for it. Let’s first recall the concept ofĀ impedance.

The impedance concept

There are three idealĀ (passive) circuit elements: the resistor, the capacitor and the inductor. RealĀ circuit elements usually combine characteristics of all of them, even if they are designed to work like ideal circuit elements. Collectively, these ideal (passive) circuit elements are referred to as impedances, because… Well… Because they have someĀ impedance. In fact, you should note that, if we reserve the terms ending with -ance for theĀ propertyĀ of the circuit elements, and those ending on -or for the objects themselves, then we should call them impedors. However, that term does not seem to have caught on.

You already know what impedance is. I explained it before, notably in my post on the intricacies related to self- and mutual inductance. Impedance basically extends the concept ofĀ resistance, as we know it fromĀ direct current (DC)Ā circuits, toĀ alternating currentĀ (AC) circuits. To put it simply, when AC currents are involved – so when the flow of charge periodically changes reverses direction – then it’s likely that, because of the properties of the circuit, the current signal will lag the voltage signal, and so we’ll have some phase differenceĀ telling us by how much. So, resistance is just a simple realĀ number R – it’sĀ the ratioĀ between (1) the voltage that is being applied across the resistor and (2) the current through it, so we write R = V/I – and it’s got aĀ magnitudeĀ only, butĀ impedance isĀ a ‘number’ that has both a magnitude as well as phase, so it’s aĀ complexĀ number, or a vector.

In engineering, such ‘numbers’ with a magnitude as well as a phase are referred to as phasors. A phasor represents voltages, currents and impedances as a phase vector (note the boldĀ italics: they explain how we got the pha-sorĀ term). It’s just a rotatingĀ vector really.Ā So a phasor has a varying magnitude (A) and phase (φ) , which is determined by (1) some maximum magnitude A0, (2) some angular frequency ω and (3) some initial phaseĀ (Īø). So we can write the amplitudeĀ A as:

A = A(φ) = A0Ā·cos(φ) =Ā A0Ā·cos(ωt + Īø)

As usual, Wikipedia has a nice animation for it:

Unfasor

In case you wonder why I am using aĀ cosineĀ rather than aĀ sineĀ function, the answer is that it doesn’t matter: the sine and the cosine are the same function except for a Ļ€/2Ā phase difference: just rotate the animation above by 90 degrees, or think about the formula: sinφ = cos(Ļ†āˆ’Ļ€/2). šŸ™‚

So A = A0Ā·cos(ωt + Īø) is the amplitude. It could be the voltage, or the current, or whatever realĀ variable.Ā The phase vectorĀ itself is represented by a complex number, i.e. a two-dimensional number, so to speak, which we can write as all of the following:

A = A0Ā·eiφ = A0Ā·cosφ + iĀ·A0Ā·sinφ = A0Ā·cos(ωt+Īø) + iĀ·A0Ā·sin(ωt+Īø)

=Ā A0Ā·ei(ωt+Īø)Ā = A0Ā·eiĪøĀ·eiωtĀ = A0Ā·eiωtĀ with A0Ā = A0Ā·eiĪø

That’s just Euler’s formula, and I am afraid I have to refer you to my page on theĀ essentialsĀ if you don’t get this. I know what you are thinking: why do we need the vector notation? Why can’t we just be happy with the A = A0Ā·cos(ωt+Īø) formula? The truthful answer is: it’s just to simplify calculations: it’s easier to work with exponentials than with cosines or sines. For example, writingĀ ei(ωt + Īø)Ā = eiĪøĀ·eiωtĀ is easier than writing cos(ωt + Īø) = … […] Well? […]Ā Hmm… šŸ™‚

See!Ā You’re stuck already. You’d have to use the cos(α+β) = cosα·cosĪ²Ā āˆ’ sinα·sinβ formula: you’d get the same results (just do it for the simple calculation of theĀ impedanceĀ below) but it takes a lot more time, and it’s easier to make mistake. Having said why complex number notation is great, I also need to warn you. There are a few things you have to watch out for. One of these things is notation. The other is the kind of mathematical operations we can do: it’s usually alright but we need to watch out with the i2Ā = –1 thing when multiplyingĀ complex numbers. However, I won’t talk about that here because it would only confuse you even more. šŸ™‚

Just for the notation, let me note that Feynman would write A0Ā as A0Ā with the little hatĀ orĀ caretĀ symbol (∧) on top of it, so as to indicate the complexĀ coefficient isĀ notĀ a variable. So he writes A0Ā asĀ Ć‚0Ā =Ā A0Ā·eiĪø.Ā However, I find that confusing and, hence, I prefer using bold-type for any complex number, variable or not. The disadvantage is that we need to remember that the coefficient in front of the exponential is not a variable: it’s a complex number alright, but not a variable. Indeed, do look at thatĀ A0Ā = A0Ā·eiĪøĀ equality carefully: A0Ā is a specificĀ complex number that captures theĀ initial phase Īø. So it’s not the magnitude of the phasor itself, i.e. |A| =Ā A0. In fact, magnitude, amplitude, phase… We’re using a lot confusing terminology here, and so that’s why you need to ‘get’ the math.

The impedance is not a variable either. It’s some constant. Having said that, this constant will depend on theĀ angular frequency ω. So… Well… Just think about this as you continue to read. šŸ™‚ So the impedance is someĀ number, just like resistance, but it’s a complexĀ number. We’llĀ denote it by ZĀ and, using Euler’s formula once again, we’ll write it as:

Z =Ā |Z|eiĪøĀ = V/I = |V|ei(ωt +Ā ĪøV)/|I|ei(ωt +Ā ĪøI)Ā = [|V|/|I|]Ā·ei(ĪøVĀ āˆ’Ā ĪøI)

So, as you can see, it is,Ā literally, some complexĀ ratio, just like R = V/I was some realĀ ratio: it is a complex ratio because itĀ has a magnitude and a direction, obviously. Also please do note that, as I mentioned already, the impedance is, in general, some function of the frequency ω, as evidenced by the ωt term in the exponential, but so we’re not looking at ω as a variable: V and I are variables and, as such, they depend on ω, but so you should look at ω as some parameter. I know I should, perhaps, notĀ be so explicit on what’s going on, but I want to make sure you understand.

So what’s going on? The illustration below (credit goes to Wikipedia, once again) explains. It’s a pretty generic view of a very simple AC circuit. So we don’t care what the impedance is: it might be an inductor or a capacitor, or a combination of both, but we don’t care: we just call it an impedance, or an impedorĀ if you want. šŸ™‚ The point is: if we apply an alternating current, then the current and the voltage will both go up and down, but the current signal will lag the voltage signal, and some phase factor Īø tells us by how much, so Īø will be the phase difference.

General_AC_circuit

Now, we’re dividing one complex number by another in that Z = V/I formula above, and dividing one complex number by another is not allĀ thatĀ straightforward, so let me re-write that formula for Z above as:

V = Iāˆ—ZĀ = Iāˆ—|Z|eiĪø

Now, while that V = Iāˆ—Z formulaĀ resembles the V = IĀ·R formula, you should note the bold-face type for V and I, and theĀ āˆ— symbol I am using here for multiplication. TheĀ bold-face for V and I implies they’re vectors, or complex numbers. As for theĀ āˆ— symbol, that’s to make it clear we’re not talking a vector cross product AƗB here, but a product of two complex numbers. [It’s obviously notĀ a vector dot product either, because a vector dot product yields a real number, not some other vector.]

Now we write V and IĀ as you’d expect us to write them:

  • VĀ =Ā |V|ei(ωt +Ā ĪøV) = V0Ā·ei(ωt +Ā ĪøV)
  • IĀ =Ā |I|ei(ωt +Ā ĪøI)Ā = I0Ā·ei(ωt +Ā ĪøI)

ĪøVĀ andĀ ĪøIĀ are, obviously, the so-called initialĀ phase of the voltage and the current respectively. These ‘initial’ phases are notĀ independent: we’re talking a phaseĀ differenceĀ really, between the voltage and the current signal, and it’s determined by the properties of the circuit. In fact, that’s the whole point here: theĀ impedanceĀ is a property of the circuit and determines how theĀ current signal varies as a function of the voltage signal. In fact, we’ll often choose the t = 0 point such that ĪøVĀ and so then we need to find ĪøI. […]Ā OK. Let’s get on with it. Writing out all of the factors in the V = Iāˆ—ZĀ = Iāˆ—|Z|eiĪøĀ equation yields:

VĀ =Ā |V|ei(ωt +Ā ĪøV)Ā =Ā Iāˆ—ZĀ = |I|ei(ωt +Ā ĪøI)āˆ—|Z|eiĪøĀ = |I||Z|ei(ωt +Ā ĪøIĀ +Ā Īø)Ā 

Now, this equation must hold for all values of t, so we can equate the magnitudes and phases and, hence, the following equalities must hold:

  1. |V| =Ā |I||Z| ⇔ |Z| =Ā |V|/|I|
  2. ωt + ĪøVĀ =  ωt + ĪøIĀ + θ ⇔ θ = ĪøVĀ āˆ’ ĪøI

Done!Ā 

Of course, you’ll complain once again about those complex numbers: voltage and current are something real, isn’t it? And so what isĀ really aboutĀ this complex numbers? Well… I can just say what I said already. You’re right. I’ve used the complex notation only to simplify the calculus, so it’s only theĀ real partĀ of those complex-valued functions that counts.

OK. We’re done with impedance. We can now discuss the impedors, including resistors (for which we won’t haveĀ such lag or phase difference, but the concept of impedance applies nevertheless).

Before I start, however, you should think about what I’ve done above: IĀ explainedĀ the concept of impedance, but I didn’t do much with it. The real-life problem will usually be that you get the voltage as a function of time, and then you’ll have to calculate the impedance of a circuit and, then, the current as a function of time. So I just showed the fundamental relations but, in real life, you won’t know what Īø andĀ ĪøIĀ could possibly be. Well… Let me correct that statement: we’ll give you formulas forĀ Īø as we discuss the various circuit elements and their impedance below, and so then you can use these formulas to calculateĀ ĪøI. šŸ™‚

Resistors

Let’s start with what seems to be the easiest thing: a resistor. A real resistor is actually not easy to understand, because it requires us to understand the properties ofĀ realĀ materials. Indeed, it may or may not surprise you, but the linear relation between the voltage and the current for realĀ materials is only approximate.Ā Also, the way resistors dissipateĀ energy is not easy to understand. Indeed, unlike inductors and capacitors, i.e. the other twoĀ passiveĀ components of an electrical circuit, a resistor does not storeĀ butĀ dissipatesĀ energy, as shown below.

Electric_load_animation_2

It’s a nice animation (credit for it has to go to Wikipedia once more), as it shows how energy is being used in an electric circuit. Note that the little moving plusesĀ are in line with the conventionĀ that a current is defined as the movement of positive charges, so we write I = dQ/dt instead ofĀ I =Ā āˆ’dQ/dt. That also explains the direction of the field line E, which has been added to show that the charges move with the field that is being generatedĀ by theĀ power source (which is not shown here). So, what we have here is that, on one side of the circuit, someĀ generator or voltage sourceĀ will create anĀ emfĀ pushing the charges,Ā and so the animation shows how someĀ load – i.e. the resistor in this case – willĀ consumeĀ their energy, so they lose their push (as shown by the change in color from yellow to black). SoĀ power, i.e.energy per unit time, is supplied, and is then consumed.

To increase the current in the circuit above, you need to increase the voltage, but increasing both amounts to increasing the powerĀ that’s being consumed in the circuit. Electric power is voltage times current, so P = VĀ·I (or vĀ·i, if I use the small letters that are used in the two animations below). Now, Ohm’s LawĀ (I = V/R)Ā says that, if we’d want to double the current, we’d need to double the voltage, and so we’re quadrupling the power then: P2Ā = V2Ā·I2Ā = (2Ā·V1)Ā·(2Ā·I1) = 4Ā·V1Ā·I1Ā = 22Ā·P1. So we have a square-cube law for the power, which we get by substituting V for RĀ·I or by substituting I for V/R, so we can write the power P as P = V2/R = I2Ā·R. This square-cube law says exactly the same: if you want to doubleĀ the voltageĀ orĀ the current, you’ll actually have to double both and, hence, you’ll quadruple the power.

But back to the impedance: Ohm’s Law is theĀ Z = V/I law for resistors, but we can simplify it because we know the voltage acrossĀ the resistorĀ and the current that’s going throughĀ areĀ in phase. Hence, ĪøVĀ andĀ ĪøIĀ are identical and, therefore, theĀ Īø = ĪøVĀ āˆ’ ĪøIĀ in Z =Ā |Z|eiĪøĀ is equal to zero and, hence, Z =Ā |Z|. Now,Ā |Z| =Ā |V|/|I| =Ā V0/I0. So the impedanceĀ ZĀ is just some real number R = V0/I0, which we can also write as:

R = V0/I0Ā = (V0Ā·ei(ωt + α))/(I0Ā·ei(ωt + α)) = V(t)/I(t), with α = ĪøVĀ = ĪøI

The equation above goes from R = V0/I0Ā to R = V(t)/I(t) = V/I. It’s note the same thing: the second equation says that,Ā at any point in time, the voltage and the current will be proportional to each other, with R or its reciprocal as the proportionality constant. In any case,Ā we have our formula for ZĀ here:

Z = R = V/I =Ā V0/I0

So that’s simple. Before we move to the next, let me note that the resistance of aĀ realĀ resistor may depend on itsĀ temperature, so in real-life applications one will want to keep its temperature as stable as possible. That’s why real-life resistors have power ratings and recommended operatingĀ temperatures. The image below illustrates how so-called heat-sink resistors can be mounted on a heat sink with a simple spring clip so as to ensureĀ the dissipated heat is transported away. These heat-sink resistors are rather small (10 by 15 mm only) but are rated for 35Ā watt – so that’s quite a lot for such small thing – ifĀ correctly mounted.

spring-clips-mounting-technigques

As mentioned, the linear relation between the voltage and the current is onlyĀ approximate, and the observed relation is also there only for frequencies that are not ‘too high’ because, if the frequency becomes very high, the free electrons will start radiating energy away, as they produce electromagnetic radiation. So one always needs to look at the tolerances of real-lifeĀ resistors, which may be ± 5%, ± 10%, or whatever.Ā In any case… On to the next.

Capacitors (condensers)

We talked at length about capacitors (aka condensers) in our post explaining capacitanceĀ or, the more widely used term, capacity: theĀ capacity of a capacitor is the observedĀ proportionality between (1) the voltage (V) across and (2) the charge (Q) on the capacitor, so we wrote it as:

C = Q/V

Now, it’s easy to confuse the C here with the C for coulomb, which I’ll also use in a moment, and so… Well… Just don’t! šŸ™‚ The meaning of the symbol is usually obvious from the context.

As for the explanation of this relation, it’s quite simple: a capacitor consists of two separateĀ conductors in space, with positive charge on one, and an equal and opposite (i.e. negative) charge on the other. Now, the logic of the superposition of fields implies that, if we double the charges, we will also double the fields, and so the work one needs to do to carry a unit charge from one conductor to the other is also doubled! So that’s why the potential difference between the conductors is proportional to the charge.

The C = Q/VĀ formula actually measures theĀ ability of the capacitor to store electric charge and, therefore, to store energy, so that’s why the term capacityĀ is really quite appropriate.Ā I’ll let youĀ googleĀ a few illustrations like the one below, that shows how a capacitor is actually being charged in a circuit. Usually, some resistance will be there in the circuit, so as to limit the current when it’s connected to the voltage source and, therefore, as you can see, the R times C factor (RĀ·C) determines how fast or how slow the capacitor charges and/or discharges. Also note that the current is equal to the time rate of change of the charge: I = dQ/dt.

images

In the above-mentioned post, we also give a few formulas for the capacity of specificĀ typesĀ of condensers. For example, for aĀ parallel-plate condenser, the formula was CĀ =  ε0A/d. We also mentioned its unit, which is isĀ coulomb/volt, obviously, but – in honor of Michael Faraday, who gave us Faraday’s Law, and many other interesting formulas – it’s referred to as theĀ farad: 1 F = 1 C/V. The C here isĀ coulomb, of course. Sorry we have to use C to denote two different things but, as I mentioned, the meaning of the symbol is usually clear from the context.

We also talked about how dielectrics actually work in that post, but we did notĀ talk about the impedance of a capacitor, so let’s do that now. The calculation is pretty straightforward. Its interpretation somewhat less so. But… Well… Let’s go for it.

It’s the current that’s charging the condenser (sorry I keep using both terms interchangeably), and we know that the current is the time rate of change of the charge (I = dQ/dt).Ā Now, you’ll remember that, in general, we’d write a phasor A as A = A0Ā·eiωtĀ with A0Ā = A0Ā·eiĪø, soĀ A0Ā is a complex coefficient incorporating the initial phase, which we wrote as ĪøVĀ andĀ ĪøIĀ for the voltage and for the current respectively. So we’ll represent the voltage and the current now using that notation, so we write: V = V0Ā·eiωtĀ and I = I0Ā·eiωt. So let’s now use that C = Q/V by re-writing it as Q = CĀ·V and, because C is some constant, we can write:

I = dQ/dt = d(CĀ·V)/dt = CĀ·dV/dt

Now, what’s dV/dt? Oh… You’ll say: V is the magnitude of V, so it’s equal toĀ |V| = |V0Ā·eiωt| = |V0|Ā·|eiωt| = |V0| = |V0Ā·eiĪø| =Ā |V0|Ā·|eiĪø| =Ā |V0| = V0. So… Well…Ā What?Ā V0Ā is some constant here! It’s theĀ maximumĀ amplitude of V, so… Well… It’s time derivative is zero: dV0/dt = 0.

Yes. Indeed. We did something very wrong here!Ā You really need to watch out with this complex-number notation, and you need to think about what you’re doing. V is not the magnitude of V but its (varying) amplitude. So it’s theĀ realĀ voltage V that varies with time: it’s equal to V0Ā·cos(ωt + ĪøV), which is the realĀ part of our phasorĀ V. Huh?Ā Yes.Ā Just hang in for a while. I know it’s difficult and, frankly, Feynman doesn’t help us very much here.Ā Let’s take one step back and so – you will see why I am doing this in a moment – let’s calculate the time derivative of our phasor V, instead of the time derivative of our realĀ voltageĀ V. So we calculate dV/dt, which is equal to:

dV/dtd(V0Ā·eiωt)/dt =Ā V0Ā·d(eiωt)/dt = V0Ā·(iω)Ā·eiωtĀ = iω·V0Ā·eiωtĀ = iω·V

Remarkable result, isn’t it? We take the time derivative of our phasor, and the result is the phasor itself multiplied with iω. Well… Yes. It’s a general property of exponentials, but still… Remarkable indeed! We’d get the same with I, but we don’t need that for the moment. What we doĀ need to do is go from ourĀ I =Ā CĀ·dV/dtĀ relation, which connects theĀ realĀ parts of I and V one to another, to theĀ I =Ā CĀ·dV/dtĀ relation, which relates theĀ (complex) phasors. So we write:

Ā I =Ā CĀ·dV/dt ⇔ I =Ā CĀ·dV/dt

Can we do that? Just like that? We just replace I and V by I and V? Yes, we can.Ā Why? Well… We know that I is the real part ofĀ I and so we can write I = Re(I)+ Im(I)Ā·i = I + Im(I)Ā·i, and then we can write the right-hand side of the equation as CĀ·dV/dt = Re(CĀ·dV/dt)+ Im(CĀ·dV/dt)Ā·i. Now, two complex numbers are equal if, and only if, their real and imaginary parts are the same, so… Well… Write it all out, if you want, using Euler’s formula, and you’ll see it all makes sense indeed.

So what do we get? TheĀ I =Ā CĀ·dV/dt gives us:

I =Ā CĀ·dV/dt = CĀ·(iω)Ā·V

That implies that I/V = CĀ·(iω) and, hence, we get – finally! – what we need to get:

Z = V/I = 1/(iωC)

This is a grand result and, while I am sorry I made you suffer for it, I think it did a good job here because, if you’d check Feynman on it, you’ll see he – or, more probably, his assistants, – just skate over this without bothering too much about mathematical rigor. OK.Ā All that’s left now is to interpret this ‘number’ Z = 1/(iωC). It is a purelyĀ imaginary number, and it’s a constant indeed, albeit a complex constant. ItĀ can be re-written as:

Z = 1/(iωC) = i-1/(ωC) = –i/(ωC) = (1/ωC)Ā·eāˆ’iĀ·Ļ€/2

[Sorry. I can’t be more explicit here. It’s just of the wonders of complex numbers: i-1 = –i. Just check one my posts on complex numbers for more detail.] Now, a –iĀ factor corresponds to a rotation ofĀ minusĀ 90 degrees, and so that gives you the true meaning of what’s usually said about a circuit with a capacitor:Ā the voltage across the capacitor willĀ lagĀ the current with a phase differenceĀ equal to π/2, as shown below. Of course, as it’s the voltage driving the current, we should say it’s the current that is lagging with a phase difference of 3Ļ€/2, rather than stating it the other way around! Indeed, i-1 = –i = –1Ā·i =Ā i2Ā·i = i3, so that amounts to three ‘turns’ of the phase in the counter-clockwise direction, which is the direction in which our ωt angle is ‘turning’.

800px-VI_phase

It is a remarkable result, though. The illustration above assumes the maximum amplitude of the voltage and the current are the same, soĀ |Z|Ā =Ā |V|/|I| = 1, but what if they are notĀ the same? What are the real bits then? I can hear you, indeed: “To hell with the bold-face letters: what’s V and I? What’s the realĀ thing?”

Well… V and I are the real bits ofĀ VĀ =Ā |V|ei(ωt+ĪøV) = V0Ā·ei(ωt+ĪøV)Ā and of IĀ =Ā |I|ei(ωt+ĪøI)Ā = I0Ā·ei(ωt+ĪøVāˆ’Īø)Ā =Ā I0Ā·ei(ωtāˆ’Īø)Ā = I0Ā·ei(ωt+Ļ€/2)Ā respectively so, assuming ĪøVĀ = 0 (as mentioned above, that’s just a matter of choosing a convenient t = 0 point), we get:

  • V = V0Ā·cos(ωt)
  • I = I0Ā·cos(ωt + Ļ€/2)

So the Ļ€/2 phase difference is there (you need to watch out with the signs, of course: Īø = āˆ’Ļ€/2, but so it’s the current that seems to leadĀ here)Ā but the V0/I0Ā ratio doesn’t have to be one, so the realĀ voltage and current could look like something below, where the maximumĀ amplitude of the current is only half of the maximumĀ amplitude of the voltage.

Capture

So let’s analyze this quickly: the V0/I0Ā ratio is equal to |Z| =Ā |V|/|I| = V0/I0Ā = 1/ωC = (1/ω)(1/C) (note that it’sĀ notĀ equal to V/I = V(t)/I(t), which is a ratio that doesn’t make sense because I(t) goes through zero as the current switches direction). So what? Well… It means the ratio is inversely proportional to both the frequency ω as well as the capacity C, as shown below. Think about this: if ω goes to zero, V0/I0Ā goes to āˆž, which means that, for a given voltage, the current must go to zero. That makes sense, because we’re talking DC current when ω → 0, and the capacitor charges itself and then that’s it: no more currents. Now, if C goes to zero, so we’re talking capacitors with hardly any capacity, we’ll also get tiny currents. Conversely, for large C, we’ll get huge currents, as the capacitor can take pretty much any charge you throw at it, so that makes for smallĀ V0/I0Ā ratios. The most interesting thing to consider is ω going to infinity, as the V0/I0Ā ratio is also quite small then. What happens? The capacitor doesn’t get the time to charge, and so it’s always in this state where it has large currents flowing in and out of it, as it can’t build the voltageĀ that would counter the electromotive force that’s being supplied by the voltage source.

graph 6OK. That’s it. Le’s discuss the last (passive) element.

Inductors

We’ve spoiled the party a bit with that illustration above, as it gives the phase difference for an inductor already:

Z = iωL = ωLĀ·eiĀ·Ļ€/2, with L the inductance of the coil

So, againĀ assuming thatĀ ĪøVĀ = 0, we can calculate I as:

IĀ =Ā |I|ei(ωt+ĪøI)Ā = I0Ā·ei(ωt+ĪøVāˆ’Īø)Ā =Ā I0Ā·ei(ωtāˆ’Īø)Ā = I0Ā·ei(ωtāˆ’Ļ€/2)Ā 

Of course, you’ll want to relate this, once again, to theĀ realĀ voltage and theĀ realĀ current, so let’s write the real parts of ourĀ phasors:

  • V = V0Ā·cos(ωt)
  • I = I0Ā·cos(ωt āˆ’ Ļ€/2)

Just to make sure you’re not falling asleep as you’re reading, I’ve made another graph of how things could look like. So now’s it’s the current signal that’sĀ lagging the voltage signalĀ with a phase differenceĀ equal toĀ Īø = Ļ€/2.

Capture

Also, to be fully complete, I should show you how the V0/I0Ā ratio now varies with L and ω. Indeed, here also we can write that |Z| =Ā |V|/|I| = V0/I0, but so here we find that V0/I0Ā =  ωL, so we have a simple linear proportionality here! For example, for a given voltage V0, we’ll have smallerĀ currents as ω increases, so that’s the opposite of what happens with our ideal capacitors. I’ll let you think about that… šŸ™‚

Capture

Now how do we get thatĀ Z = iωL formula? In my post on inductance, I explained what an inductor is: a coil of wire, basically. Its defining characteristic is thatĀ a changing current will cause a changingĀ magnetic field in it and, hence, some change in the flux of the magnetic field. Now, Faraday’s Law tells us that that will cause some circulation of the electric field in the coil, which amounts to an inducedĀ potential difference which is referred to as theĀ electromotive forceĀ (emf). Now, it turns out that the inducedĀ emfĀ is proportional to the change in current. So we’ve got another constant of proportionality here, so it’s like how we defined resistance, orĀ capacitance. So, in many ways, theĀ inductanceĀ is just another proportionality coefficient. If we denote it by L – the symbol is said to honor the Russian phyicistĀ Heinrich Lenz, whom you know from Lenz’ Law – then we define it as:

L = āˆ’Ę/(dI/dt)

The dI/dt factor is, obviously, the time rate of change of the current, and the negative sign indicates that the emfĀ opposesĀ the change in current, so it will tend to cause anĀ opposingĀ current. However, theĀ powerĀ of our voltage source will ensure the current does effectively change, so it will counter the ‘back emf’ that’s being generated by the inductor. To be precise, the voltage across the terminals of our inductor, which we denote by V, will be equal and opposite to ʐ, so we write:

V = āˆ’Ę = LĀ·(dI/dt)

Now, this very much resembles the I =Ā CĀ·dV/dt relation we had for capacitors, and it’s completely analogous indeed: we just need to switch the I and V, and C and L symbols. So we write:

Ā V = LĀ·dI/dt⇔ V = LĀ·dI/dt

Now, dI/dt is a similar time derivative as dV/dt. We calculate it as:

dI/dtd(I0Ā·eiωt)/dt =Ā I0Ā·d(eiωt)/dt = I0Ā·(iω)Ā·eiωtĀ = iω·I0Ā·eiωtĀ = iω·I

So we get what we want and have to get:

V = LĀ·dI/dt = iωLĀ·I

Now, Z = V/I, so Z =Ā iωL indeed!

Summary of conclusions

Let’s summarize what we found:

  1. For a resistor, we have Z(resistor) = ZRĀ = R = V/I =Ā V0/I0
  2. For an capacitor, we have Z(capacitor) = ZCĀ = 1/(iωC) = –i/(ωC)
  3. For an inductor, we haveĀ Z(inductance) = ZL= iωL

Note that the impedance of capacitors decreases as frequency increases, while for inductors, it’s the other way around. We explained that by making you think of the currents: for a given voltage, we’ll have large currents for high frequencies, and, hence, a smallĀ V0/I0Ā ratio. Can you think of what happens with an inductor? It’sĀ notĀ so easy, so I’ll refer you to the addendum below for some more explanation.

Let me also note that, as you can see, the impedance of (ideal) inductors and capacitors is a pure imaginary number, so that’s a complex number which has no real part. In engineering, the imaginary part of the impedance is referred to as theĀ reactance, so engineers will say that ideal capacitors and inductors have aĀ purely imaginary reactive impedance.Ā 

However, in real life, the impedance will usually have both a real as well as an imaginary part, so it will be some kind of mix, so to speak. The real part is referred to as the ‘resistance’ R, and the ‘imaginary’ part is referred to as the ‘reactance’ X. The formula for both is given below:

formula resistance and reactance

But here I have to end my post on circuit elements. It’s become quite long, so I’ll discuss Kirchoff’s rulesĀ in my next post.

Addendum: Why is V = āˆ’ ʐ?

Inductors are not easy to understand—intuitively, that is. That’s why I spent so much time writing on them in my other post on them, to which I should be referring you here. But let me recapitulate the key points. The key idea is that we’re pumping energy into an inductor when applying a current and, as you know, the time rate of change is power: P = dW/dt, so we’re talking power here too, which is voltageĀ timesĀ current: P = dW/dt = VĀ·I.Ā The illustration below shows what happens when anĀ alternating currentĀ is applied to the circuit with the inductor. So the assumption is that the current goes in one and then in the other direction, so I > 0, and then I < 0, etcetera. We’re also assuming some nice sinusoidal curve for the current here (i.e. the blue curve), and so we get what we get for U (i.e. the red curve), which is the energy that’s stored in the inductor really, as it tries to resist the changing current: the energy goes up and down between zero and some maximum amplitude that’s determined by the maximum current.

power 2

So, yes, building up current requires energy from some external source, which is used to overcome the ‘back emf’ in the inductor, and that energy is stored in the inductor itself. [If you still wonder why it’s stored in the inductor, think about the other question: where else would it be stored?]Ā HowĀ is stored? Look at the graph and think: it’s stored as kinetic energy of the charges, obviously. That explains why the energy is zero when the current is zero, and why the energy maxes out when the current maxes out. So, yes, it all makes sense! šŸ™‚

Let me give another example.Ā The graph below assumes the current builds up to some maximum. As it reaches its maximum, theĀ storedĀ energy will also max out. This example assumesĀ direct current, so it’s a DC circuit: the current builds up, but then stabilizes at some maximum that we can find by applying Ohm’s Law to the resistanceĀ of the circuit: I = V/R. Resistance? But we were talking an ideal inductor? We are. If there’s no other resistance in the circuit, we’ll have a short-circuit, so the assumption is that we doĀ have some resistance in the circuit and, therefore, we should also think of some energyĀ lossĀ to heat from the current in the resistance. If not, well… Your power source will obviously soon reach its limits. šŸ™‚

power

So what’s going on then?Ā We have someĀ changingĀ current in the coil but, obviously, some kind of inertia also: the coil itself opposes the change in current through the ‘back emf’. Now, it requires energy, or power, to overcome the inertia, so that’s the power that comes from our voltage source: it will offset the ‘back emf’, so we may effectively think of a little circuit with an inductorĀ and a voltage source, as shown below.

circuit with coil

But why do we write V = āˆ’Ā Ę? Our voltage source can have any voltage, can’t it? Yes. Sure. But so the coil will always provide anĀ emf that’s exactly theĀ oppositeĀ of this voltage. Think of it: we have someĀ voltage that’s being applied across the terminals of the inductor, and so we’ll have some current. A current that’s changing. And it’s that current will generate an emf that’s equal to Ɛ = –LĀ·(dI/dt). So don’t think of ʐ as some constant: it’s the self-inductance coefficient L that’s constant, but I (and, hence, dI/dt) and V are variable.

The point is: we cannot have any potential difference in a perfect conductor, which is what the terminals are: any potential difference, i.e. any electric field really, would cause huge currents. In other words, the voltage V and the emf Ɛ have to cancel each other out,Ā all of the time. If not, we’d have huge currents in the wires re-establishing the V = āˆ’Ę equality.

Let me use Feynman’s argument here. Perhaps that will work better. šŸ™‚ Our ideal inductor is shown below: it’s shielded by some metal box so as to ensure it does not interact with the rest of the circuit. So we have some current I, which we assume to be an AC current, and we know someĀ voltageĀ is needed toĀ causeĀ that current, so that’s the potential difference V between the terminals.

inductor

The total circulation of E – around the whole circuit – can be written as the sum of two parts:

Formula circulaton

Now, we know circulation of E can only be caused by someĀ changingĀ magnetic field, which is what’s going on in the inductor:

emf

So this change in the magnetic flux is what it causing the ‘back emf’, and so the integral on the left is, effectively, equal to ʐ, not minus Ɛ but +ʐ.Ā Now, the second integral is equal to V, because that’s the voltage V between the two terminals a and b. So the whole integral is equal toĀ 0 = ʐ +Ā V and, therefore, we have that:

VĀ = āˆ’Ā Ę = LĀ·dI/dt

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/

Re-visiting the speed of light, Planck’s constant, and the fine-structure constant

Note: I have published a paper that is very coherent and fully explains what the fine-structure constant actually is. There is nothing magical about it. It’s not some God-given number. It’s a scaling constant – and then some more. But not God-given. Check it out: The Meaning of the Fine-Structure Constant. No ambiguity. No hocus-pocus.

Jean Louis Van Belle, 23 December 2018

Original post:

A brother of mine sent me a link to an article he liked. Now, because we share some interest in physics and math and other stuff, I looked at it and…

Well… I was disappointed. Despite the impressive credentials of its author – a retired physics professor – it was veryĀ poorly written. It made me realize how much badly written stuff is around, and I am glad I am no longer wasting my time on it. However, I do owe my brother some explanation of (a) why I think it was bad, and of (b) what, in my humble opinion, he should be wasting his time on. šŸ™‚ So what it is all about?

The article talks about physicists deriving the speed of light from “theĀ electromagnetic properties of the quantum vacuum.” Now, it’s theĀ term ‘quantum‘, in ‘quantum vacuum’, that made me read the article.

Indeed, derivingĀ the theoretical speed of light in empty space from the properties of theĀ classicalĀ vacuum – aka empty space – is a piece of cake: it was done by Maxwell himself as he was figuring out his equations back in the 1850s (see my post on Maxwell’s equations and the speed of light). And then he compared it to the measuredĀ value, and he saw it was right on the mark. Therefore, saying that the speed of light is a property of the vacuum, or of empty space, is like a tautology: we may just as well put it the other way around, and say that it’s the speed of light thatĀ definesĀ the (properties of the) vacuum!

Indeed, as I’ll explain in a moment: the speed of lightĀ determinesĀ both the electric as well as the magnetic constants μ0Ā and ε0, which are the (magnetic) permeability and the (electric) permittivity of the vacuum respectively. Both constants depend on the units we are working with (i.e. the units for electric charge, for distance, for time and for force – or for inertia, if you want, because force is defined in terms of overcoming inertia), but so they are just proportionality coefficients in Maxwell’s equations. So once we decide what units to use in Maxwell’s equations, then μ0Ā and ε0Ā areĀ just proportionality coefficientsĀ which we get from c. So they areĀ notĀ separate constants really – I mean, they are not separate from c – and all of the ‘properties’ of the vacuum, including these constants, are in Maxwell’s equations.

In fact, when Maxwell compared the theoreticalĀ value ofĀ cĀ with its presumed actual value, he didn’t compare c‘s theoretical valueĀ with the speed of light as measured by astronomers (like that 17th century Ole Roemer, to which our professor refers: he had a first go at it by suggesting some specific value for it based on his observations of the timing of the eclipses of one of Jupiter’s moons), but with c‘s value as calculated from the experimentalĀ values of μ0Ā and ε0!Ā So he knew veryĀ well what he was looking at. In fact, to drive home the point, it may also be useful to note that the Michelson-Morley experiment – which accuratelyĀ measured the speed of light – was done some thirty years later. So Maxwell had already left this world by then—very much in peace, because he had solved the mystery all 19th century physicists wanted to solve through his greatĀ unification: his set of equations covers it all, indeed: electricity, magnetism, light, and even relativity!

I think the article my brother likedĀ so much does a very lousy job in pointing all of that out, but that’s not why I wouldn’tĀ recommend it.Ā It got my attention because I wondered why one would try to derive the speed of light from the properties of the quantum vacuum. In fact, to be precise, I hoped the article would tell me what the quantumĀ vacuum actually is. Indeed, as far as I know, there’s only one vacuum—one ’empty space’: empty is empty, isn’t it? šŸ™‚ So I wondered: do we have a ‘quantum’ vacuum? And, if so, what is it, really?

Now, thatĀ is where the article is reallyĀ disappointing, I think. The professor dropsĀ a few names (like the Max Planck Institute, the University of Paris-Sud, etcetera), and then, promisingly, mentions ‘fleeting excitations of the quantum vacuum’ and ‘virtual pairs of particles’, but then he basically stops talking about quantum physics. Instead, he wanders off to share some philosophical thoughts on the fundamental physical constants. What makes it all worse is that even thoseĀ thoughts on the ‘essential’ constants are quite off the mark.

So… This post is just a ‘quick and dirty’ thing for my brother which, I hope, will be somewhat more thought-provoking than that article. More importantly, I hope that my thoughts will encourage him to try to grind through better stuff.

On Maxwell’s equations and the properties of empty space

Let me first say something about the speed of light indeed. Maxwell’s four equations mayĀ look fairlyĀ simple, but that’s only until one starts unpacking all those differential vector equations, and it’s only when going through all of their consequences that one starts appreciating their deep mathematical structure. Let me quickly copy how another bloggerĀ jotted them down: šŸ™‚

god-said-maxwell-equation

As I showed in my above-mentioned post, the speed of lightĀ (i.e. the speed with which an electromagnetic pulse or wave travels through space)Ā is just one of the many consequences of the mathematical structure of Maxwell’s set of equations. As such, the speed of light is a direct consequence of the ‘condition’, or the properties, of the vacuum indeed, as Maxwell suggested when he wrote thatĀ ā€œwe can scarcely avoid the inference that light consists in the transverse undulations of the same medium which is the cause of electric and magnetic phenomenaā€.

Of course, whileĀ Maxwell still suggests light needs some ‘medium’ here – so that’s a reference to the infamous aetherĀ theory – we now know that’s because he was a 19th century scientist, and so we’ve done away with the aether concept (because it’s a redundant hypothesis), and so now we also know there’s absolutely no reason whatsoever to try to “avoid the inference.” šŸ™‚ It’s all OK, indeed: light is some kind of “transverse undulation” of… Well… Of what?

We analyze light as traveling fields, represented by two vectors, E and B, whose direction and magnitude varies both in space as well as in time.Ā E and BĀ are field vectors, and represent the electric and magnetic field respectively. An equivalent formulation – more or less, that is (see my post on theĀ LiĆ©nard-Wiechert potentials) – for Maxwell’s equations when only one (moving) charge is involved is:

E

B

This re-formulation, which is Feynman’s preferred formula for electromagnetic radiation, is interesting in a number of ways. It clearly shows that, while we analyze the electric and magnetic field as separateĀ mathematicalĀ entities, they’re one and the same phenomenon really, as evidenced by the B = –erā€˜Ć—E/cĀ equation, which tells us the magnetic field from a single moving charge is always normal (i.e. perpendicular) to the electric field vector, and also that B‘s magnitude is 1/cĀ times the magnitude of E, soĀ |B| = B =Ā |E|/c = E/c. In short, B is fully determined by E, or vice versa: if we have one of the two fields, we have the other, so they’re ‘one and the same thing’ really—not in a mathematical sense, but in a real sense.

Also note that E and B‘s magnitude is just the same if we’re using natural units, so if we equateĀ cĀ with 1. Finally, as I pointed out in my post on the relativity of electromagnetic fields, if we would switch from one reference frame to another, we’ll have a different mix of E and B, but that different mix obviously describes the sameĀ physicalĀ reality. More in particular, if we’d be movingĀ with the charges, the magnetic field sort of disappears to re-appear as an electric field. So the Lorentz forceĀ F = FelectricĀ +Ā FmagneticĀ = qE + qvƗBĀ is one force really, and its ‘electric’ and ‘magnetic’ component appear the way they appear in our reference frame only. In some other reference frame, we’d have the same force, but its components would look different, even if they, obviously, would and should add up to the same.Ā [Well… Yes and no… You know there’s relativistic corrections to be made to the forcesĀ to, but that’s a minor point, really. The force surely doesn’t disappear!]

All of this reinforces what you know already: electricity and magnetism are part and parcel of one and the same phenomenon,Ā the electromagnetic force field, and Maxwell’s equations are the mostĀ elegant way of ‘cutting it up’. Why elegant? Well… Click the Occam tab. šŸ™‚

Now, after having praised Maxwell once more, I must say that Feynman’s equations above have another advantage. In Maxwell’s equations, we see two constants, the electric and magnetic constant (denoted by μ0Ā and ε0Ā respectively), andĀ Maxwell’s equations imply that the productĀ of theĀ electric and magnetic constant is the reciprocal of c2: μ0·ε0Ā = 1/c2. So here we see ε0Ā and cĀ only, so no μ0, so that makes it even more obvious that the magnetic and electric constant are related one to another through c.

[…] Let me digress briefly: why do we have c2Ā in μ0·ε0Ā = 1/c2, instead of justĀ c? That’s related to the relativistic nature of the magnetic force: think about that B = E/c relation. Or, better still, think about the Lorentz equation F = FelectricĀ +Ā FmagneticĀ = qE + qvƗB = q[EĀ + (v/c)Ɨ(EƗerā€˜)]: the 1/c factor is there because the magnetic force involves some velocity, and any velocity is always relative—and here I don’t mean relative to the frame of reference but relative to the (absolute) speed of light! Indeed, it’s the v/c ratio (usually denoted by β = v/c) that enters all relativistic formulas. So the left-hand side of the μ0·ε0Ā = 1/c2Ā equation is best written as (1/c)Ā·(1/c), with one of the two 1/c factors accounting for the fact that the ‘magnetic’ force is a relativistic effect of the ‘electric’ force, really, and the other 1/cĀ factor giving us the proper relationship between the magnetic and the electric constant.Ā To drive home the point, I invite you to think about the following:

  • μ0Ā is expressed in (VĀ·s)/(AĀ·m), while ε0Ā is expressed in (AĀ·s)/(VĀ·m), so the dimension in which the μ0·ε0Ā product is expressed is [(VĀ·s)/(AĀ·m)]Ā·[(AĀ·s)/(VĀ·m)] = s2/m2, so that’s the dimension of 1/c2.
  • Now, this dimensional analysis makes it clear that we can sort of distributeĀ 1/c2Ā over the two constants. All it takes is re-defining the fundamental unitsĀ we use to calculate stuff, i.e. the units for electric charge, for distance, for time and for force – or for inertia, as explained above. But so we could, if we wanted, equate both μ0Ā as well as ε0Ā with 1/c.
  • Now, if we would then equate c with 1, we’d have μ0Ā = ε0Ā = cĀ = 1. We’d have to define our units for electric charge, for distance, for time and for force accordingly, but it could be done, and then we could re-write Maxwell’s set of equations using these ‘natural’ units.

In any case, the nitty-gritty here is less important: the point is that μ0Ā and ε0Ā are also related through the speed of light and, hence, they are ‘properties’ of the vacuum as well. [I may add that this is quite obvious if you look at their definition, but we’re approaching the matter from another angle here.]

In any case, we’re done with this. On to the next!

On quantum oscillations, Planck’s constant, and Planck unitsĀ 

The second thought I want to develop is about the mentioned quantum oscillation. What is it? Or what could it be? An electromagnetic wave is caused by a moving electric charge. What kind of movement? Whatever: the charge could move up or down, or it could just spin around some axis—whatever, really. For example, if it spins around some axis, it will have a magnetic moment and, hence, the field is essentially magnetic, but then, again,Ā E and B are related and so it doesn’t really matter if the first cause is magnetic or electric: that’s just our way of looking at the world: in another reference frame, one that’s moving with the charges,Ā the field would essential be electric.Ā So the motion can be anything: linear, rotational, or non-linear in some irregular way. It doesn’t matter: any motion canĀ always be analyzed as the sum of a number of ‘ideal’ motions.Ā So let’s assume we have some elementaryĀ charge in space, and it moves and so it emits some electromagnetic radiation.

So nowĀ we need to think about that oscillation. The key question is: how small can it be? Indeed, in one of my previous posts, I tried to explain some of the thinking behind the idea of the ‘Great Desert’, as physicists call it. The whole idea is based on our thinking about the limit: what is the smallest wavelength that still makes sense? So let’s pick up that conversation once again.

The Great Desert lies between theĀ 1032Ā and 1043 Hz scale. 1032Ā Hz corresponds to aĀ photonĀ energy of Eγ = hĀ·f = (4Ɨ10āˆ’15Ā eVĀ·s)Ā·(1032Ā Hz) = 4Ɨ1017Ā eV = 400,000 tera-electronvolt (1 TeV = 1012Ā eV). I use the γ (gamma) subscript in my Eγ symbol for two reasons: (1) to make it clear that I am not talking the electric field E here but energy, and (2) to make it clear we are talking ultra-high-energy gamma-rays here.

In fact, γ-rays of this frequency and energy are theoretical only. Ultra-high-energy gamma-rays are defined as rays with photon energies higher than 100 TeV, which is the upper limit forĀ very-high-energy gamma-rays, which have been observed as part of the radiation emitted by so-called gamma-ray burstsĀ (GRBs):Ā flashes associated with extremely energetic explosions in distant galaxies. Wikipedia refers to them as the ‘brightest’ electromagnetic events know to occur in the Universe. These rays are not to be confused withĀ cosmicĀ rays, which consist of high-energy protonsĀ andĀ atomic nuclei stripped of their electron shells. Cosmic rays aren’t rays really and, because they consist of particles with a considerable rest mass, their energy is even higher. The so-called Oh-My-GodĀ particle, for example, which is the most energetic particle ever detected, had an energy ofĀ 3Ɨ1020Ā eV, i.e.Ā 300 million TeV. But it’s not a photon: its energy is largely kinetic energy, with the rest mass m0Ā counting for a lot in the mĀ in the E =Ā mĀ·c2Ā formula. To be precise: the mentioned particle was thought to be anĀ ironĀ nucleus, and it packed the equivalent energy of a baseball traveling at 100 km/h!Ā 

But let me refer you to another source for a good discussion on these high-energy particles, so I can get get back to the energy of electromagnetic radiation. When I talked about the Great Desert in that post, I did so using the Planck-Einstein relation (E = hĀ·f), which embodies the idea of the photon being valid always and everywhere and, importantly,Ā at every scale. I also discussed the Great Desert using real-life light being emitted by real-life atomic oscillators. Hence, I may have given the (wrong) impression that the idea of a photon as a ā€˜wave train’ is inextricably linked with these real-life atomic oscillators, i.e. to electrons going from one energy level to the next in some atom. Let’s explore these assumptions somewhat more.

Let’s start with the second point. Electromagnetic radiation is emitted by any accelerating electric charge, so the atomic oscillator model is an assumption that shouldĀ notĀ be essential. And it isn’t. For example, whatever is left of the nucleus after alpha or beta decay (i.e. a nuclear decayĀ process resulting in the emission of an α- or β-particle) it likely to be in an excited state, and likely to emit a gamma-ray for about 10āˆ’12 seconds, so that’s a burst that’s about 10,000 times shorterĀ than the 10–8Ā seconds it takes for the energy of a radiating atom to die out. [As for the calculation of that 10–8Ā sec decay time – so that’s like 10 nanoseconds – I’ve talked about this before but it’s probably better to refer you to the source, i.e. one of Feynman’s Lectures.]

However, what we’re interested in is not the energy of theĀ photon, but the energy of one cycle. In other words, we’re not thinking of the photon as some wave train here, but what we’re thinking about is the energy that’s packed into a space corresponding toĀ one wavelength. What can we say about that?

As you know, that energy will depend both on the amplitudeĀ of the electromagnetic wave as well as its frequency. To be precise, the energy is (1) proportional to theĀ squareĀ of the amplitude, and (2) proportional to the frequency. Let’s look at the first proportionality relation. It can be written in a number of ways, but one way of doing it is stating the following: if we know the electric field, then the amount of energy that passes per square meter per second through a surface that is normal to the direction in which the radiation is going (which we’ll denote by S – the s from surface – in the formula below), must be proportional to the average of the square of the field. So we write SĀ āˆĀ āŒ©E2〉, and so we should think about theĀ constant of proportionalityĀ now. Now, let’s not get into the nitty-gritty, and so I’ll just referĀ to Feynman for the derivation of the formula below:

S = ε0c·〈E2〉

So the constant of proportionality is ε0c. [Note that, in light of what we wrote above, we can also write this asĀ S = (1/μ0Ā·c)·〈(cĀ·B)2〉 = (c/μ0)·〈B2〉, so that underlines once again that we’re talking one electromagneticĀ phenomenon only really.] So that’sĀ a nice and rather intuitive result in light of all of the other formulas we’ve been jotting down. However, it isĀ a ‘wave’ perspective. The ‘photon’ perspective assumes that, somehow, the amplitude is given and, therefore, the Planck-Einstein relation only captures the frequency variable:Ā Eγ = hĀ·f.

Indeed, ‘more energy’ in the ‘wave’ perspective basically means ‘more photons’, but photons are photons: they have a definiteĀ frequencyĀ and a definiteĀ energy, and both are given by that Planck-Einstein relation.Ā So let’s look at that relation by doing a bit of dimensionalĀ analysis:

  • Energy is measured in electronvolt or, using SI units, joule: 1 eV ā‰ˆĀ 1.6Ɨ10āˆ’19Ā J. Energy is force times distance: 1 joule = 1 newtonĀ·meter, which means that a larger force over a shorter distance yields the same energy as a smaller force over a longer distance. The oscillations we’re talking about here involve very tinyĀ distances obviously. But the principle is the same: we’re talking some moving chargeĀ q, and theĀ power – which is the time rate of change of the energy – that goes in or out at any point of time is equal to dW/dt =Ā FĀ·v, with W the work that’s being done by the charge as it emits radiation.
  • I would also like to add that, as you know, forces are related to theĀ inertiaĀ of things. Newton’s Law basically defines a force as that what causes a mass to accelerate: F = mĀ·a =Ā mĀ·(dv/dt) = d(mĀ·v)/dt = dp/dt, with p theĀ momentumĀ of the object that’s involved. When charges are involved, we’ve got the same thing: aĀ potential differenceĀ will cause some currentĀ to change, and one of the equivalents of Newton’s LawĀ F = mĀ·a = mĀ·(dv/dt) in electromagnetism is V = LĀ·(dI/dt). [I am just saying this so you get a better ‘feel’ for what’s going on.]
  • Planck’s constant is measured inĀ electronvoltĀ·secondsĀ (eVĀ·s) or in, using SI units, inĀ jouleĀ·secondsĀ (JĀ·s), so its dimension is that of (physical) action, which is energy times time: [energy]Ā·[time]. Again, a lot of energy during a short time yields the same energy as less energy over a longer time. [Again, I am just saying this so you get a better ‘feel’ for these dimensions.]
  • The frequency f is the number of cycles per time unit, so that’s expressedĀ per second, i.e. inĀ herzĀ (Hz) = 1/second = sāˆ’1.

So… Well… It all makes sense: [xĀ joule]Ā = [6.626Ɨ10āˆ’34Ā joule]Ā·[1 second]Ɨ[fĀ cycles]/[1 second].Ā But let’s try to deepen our understanding even more: what’s the Planck-Einstein relationĀ reallyĀ about?

To answer that question, let’s think some more about the wave function. As you know, it’s customary to express the frequency as an angular frequency ω, as used in the wave function A(x, t) = A0Ā·sin(kx āˆ’ ωt).Ā The angular frequency is the frequency expressed inĀ radiansĀ per second. That’s because we need an angleĀ in our wave function, and so we need to relate x and t to some angle.Ā The way to think about this is as follows: one cycle takes a time T (i.e. the periodĀ of the wave)Ā which is equal to T = 1/f. Yes: one second divided by the number of cycles per second gives you the time that’s needed for one cycle. One cycle is also equivalent to our argument ωt going around the full circle (i.e. 2Ļ€), so we write:Ā  ω·T = 2Ļ€ and, therefore:

ω = 2Ļ€/T = 2π·f

Now we’re ready to play with the Planck-Einstein relation. WeĀ knowĀ it gives us the energy ofĀ oneĀ photon really, but what ifĀ we re-write our equation Eγ = hĀ·fĀ as Eγ/fĀ = h? The dimensions in this equation are:

[xĀ joule]Ā·[1 second]/[fĀ cyles] = [6.626Ɨ10āˆ’34Ā joule]Ā·[1 second]

⇔ xĀ =Ā 6.626Ɨ10āˆ’34Ā jouleĀ per cycle

So that means that the energy per cycle is equal toĀ 6.626Ɨ10āˆ’34Ā joule,Ā i.e. the value of Planck’s constant.

Let me rephrase trulyĀ amazing result, so you appreciate it—perhaps: regardless of the frequency of the light (or our electromagnetic wave, in general) involved, the energy per cycle, i.e. per wavelength or per period, is alwaysĀ equal to 6.626Ɨ10āˆ’34Ā joule or, using the electronvolt as the unit, 4.135667662Ɨ10āˆ’15Ā eV. So, in case you wondered,Ā thatĀ is the trueĀ meaning of Planck’s constant!

Now, if we have the frequency f, we also have the wavelength Ī», because the velocity of the wave is the frequency times the wavelength:Ā cĀ = λ·f and, therefore, Ī» = c/f.Ā So if we increase the frequency, the wavelength becomes smaller and smaller, and so we’re packing the same amount of energy – admittedly, 4.135667662Ɨ10āˆ’15Ā eV is a very tiny amount of energy – into a space that becomes smaller and smaller. Well… What’s tiny, and what’s small? All is relative, of course. šŸ™‚ So that’s where the Planck scale comes in. IfĀ we pack that amount of energy into some tiny little space of the Planck dimension, i.e. a ā€˜length’ of 1.6162Ɨ10āˆ’35Ā m, then it becomes a tiny black hole, and it’s hard to think about how that would work.

[…] Let me make a small digression here. I said it’s hard to think about black holes but, of course, it’s not because it’s ‘hard’ that we shouldn’t try it. So let me just mention a few basic facts. For starters, black holes do emit radiation! So they swallow stuff, but they also spit stuff out. More in particular, there is the so-called Hawking radiation,Ā as Roger Penrose and Stephen Hawking discovered.

Let me quickly make a few remarks on that: Hawking radiation is basically a form of blackbody radiation, so all frequencies are there, as shown below: theĀ distributionĀ of the various frequencies depends on the temperature of the black body, i.e. the black hole in this case. [The black curve is the curve that Lord Rayleigh and Sir James Jeans derived in the late 19th century, usingĀ classicalĀ theory only, so that’s the one that does notĀ correspond to experimental fact, and which led Max Planck to become the ‘reluctant’ father of quantum mechanics. In any case, that’s history and so I shouldn’t dwell on this.]

600px-Black_body

The interesting thing about blackbody radiation, includingĀ Hawking radiation, is that it reduces energy and, hence, the equivalent mass of our blackbody. SoĀ Hawking radiation reduces the mass and energy of black holes and is therefore also known as black hole evaporation. So black holes that lose more mass than they gain through other means are expected to shrink and ultimately vanish. Therefore, there’s all kind of theories that say why micro black holes, like that Planck scale black hole we’re thinking of right now, should be muchĀ larger net emitters of radiation than large black holes and, hence, whey they should shrink and dissipate faster.

Hmm… Interesting… What do we do with all of this information? Well… Let’s think about it as we continue our trek on this long journey to reality over the next year or, more probably, years (plural). šŸ™‚

The key lesson here is that space and time are intimately related because of the idea of movement, i.e. the idea of something having some velocity, and that it’s not so easy to separate the dimensions of time and distance in any hard and fast way. As energy scales become larger and, therefore, our natural time and distance units become smaller and smaller, it’s the energy concept that comes to the fore.Ā It sort of ‘swallows’ all other dimensions, and it does lead to limiting situations which are hard to imagine. Of course, that just underscores the underlying unity of Nature, and the mysteries involved.

So… To relate all of this back to the story that our professor is trying to tell, it’s a simple story really. He’s talking aboutĀ two fundamental constants basically, c and h, pointing out thatĀ cĀ is a property of empty space, andĀ h is related to something doing something. Well… OK. That’s really nothing new, and surely notĀ ground-breaking research. šŸ™‚

Now, let me finish my thoughts on all of the above by making one more remark. If you’ve read a thing or two about this – which you surely have – you’ll probably say: this is not how people usually explain it. That’s true, they don’t. Anything I’ve seen about this just associates theĀ 1043 Hz scale with the 1028Ā eV energy scale, using the same Planck-Einstein relation. For example, the Wikipedia article on micro black holes writes that “the minimum energy of a microscopic black hole is 1019Ā GeV [i.e. 1028Ā eV], which would have to be condensed into a region on the order of the Planck length.” So that’s wrong. I want to emphasize this point because I’ve been led astray by it for years.Ā It’s not the totalĀ photon energy, butĀ the energy per cycle that counts. Having said that, it is correct, however, and easy to verify, that theĀ 1043 Hz scale corresponds to a wavelength of the Planck scale: Ī» =Ā c/fĀ = (3Ɨ108Ā m/s)/(1043Ā sāˆ’1) = 3Ɨ10āˆ’35Ā m. The confusion between the photon energy and the energy per wavelengthĀ arises because of the idea of a photon: it travels at the speed of light and, hence, because of theĀ relativistic length contraction effect, it is said to be point-like, to have no dimension whatsoever. So that’s why we think of packing all of its energy in some infinitesimally small place. But you shouldn’t think like that. The photon is dimensionless in ourĀ reference frame: in its own ‘world’, it is spread out, so itĀ isĀ a wave train. And it’s in its ‘own world’ that the contradictions start… šŸ™‚

OK. Done!

My third and final point is about what our professor writes on the fundamental physical constants, and more in particular on what he writes on the fine-structure constant. In fact, I could just refer you toĀ my own post on it, but that’s probably a bit too easy for me and a bit difficult for you šŸ™‚ so let me summarize that post and tell you what you need to know about it.

The fine-structure constant

The fine-structure constant α is a dimensionlessĀ constant whichĀ also illustrates the underlying unity of Nature, but in a way that’sĀ much more fascinatingĀ than the two or three things the professor mentions. Indeed, it’s quite incredible how this number (α =Ā 0.00729735…, but you’ll usually see it written as its reciprocal, which is a number that’s close to 137.036…) links charge with the relative speeds, radii, and the mass of fundamental particles and, therefore, how this number also these concepts with each other. And, yes, the fact that it is, effectively, dimensionless, unlike h or c, makes it even more special.Ā Let me quickly sum up what the very same number α all stands for:

(1) α is the square of the electron charge expressed in Planck units: α = eP2.

(2) α is the square root of the ratio of (a) the classical electron radius and (b) the Bohr radius: α = √(reĀ /r). You’ll see this more often written asĀ reĀ = α2r. Also note that this is an equation that doesĀ notĀ depend on the units, in contrast to equation 1 (above), and 4 and 5 (below), which require you to switch to Planck units. It’s the square of a ratio and, hence, the units don’t matter. They fall away.

(3) α is the (relative) speed of an electron: α = v/c. [The relative speed is the speed as measured against the speed of light. Note that the ā€˜natural’ unit of speed in the Planck system of units is equal to c. Indeed, if you divide one Planck length by one Planck time unit, you get (1.616Ɨ10āˆ’35Ā m)/(5.391Ɨ10āˆ’44Ā s) =Ā cĀ m/s. However, this is another equation, just like (2), that does notĀ depend on the units: we can express v and c in whatever unit we want, as long we’re consistent and express both in the same units.]

(4) α is also equal to the product of (a) the electron mass (which I’ll simply write as meĀ here) and (b) the classical electron radius reĀ (if both are expressed in Planck units): α =Ā meĀ·re. Now IĀ thinkĀ that’s, perhaps, theĀ mostĀ amazing of all of the expressions for α. [If you don’t think that’s amazing, I’d really suggest you stop trying to study physics. :-)]

Also note that, from (2) and (4), we find that:

(5) The electron mass (in Planck units) is equal me = α/re = α/α2r = 1/αr. So that gives us an expression, using α once again, for the electron mass as a function of the Bohr radius r expressed in Planck units.

Finally, we can also substitute (1) in (5) to get:

(6) The electron mass (in Planck units) is equal to me = α/re  = eP2/re. Using the Bohr radius, we get me = 1/αr = 1/eP2r.

So… As you can see, this fine-structure constant really links allĀ of the fundamental properties of the electron: its charge, its radius, its distance to the nucleus (i.e. the Bohr radius), its velocity, its mass (and, hence, its energy),…

So… Why is what it is?

Well… We all marvel at this, but what can we say about it, really? I struggle how to interpret this, just as much – or probably much more šŸ™‚ – as the professor who wrote the article I don’t like (because it’s so imprecise, and that’s what made me write all what I am writing here).

Having said that, it’s obvious that it points toĀ a unity beyond these numbers and constants that I am only beginningĀ to appreciate for what it is: deep, mysterious, and very beautiful.Ā But so I don’t think that professor does a good job at showing how deep, mysterious and beautiful it all is. But then that’s up to you, my brother and you, my imaginary reader, to judge, of course. šŸ™‚

[…] I forgot to mention what I mean with ‘Planck units’. Well… Once again, I should refer you to one of my other posts. But, yes, that’s too easy for me and a bit difficult for you. šŸ™‚ So let me just note we get those Planck units by equating not less thanĀ five fundamental physical constants to 1, notably (1) the speed of light, (2) Planck’s (reduced) constant, (3) Boltzmann’s constant, (4) Coulomb’s constant and (5) Newton’s constant (i.e. the gravitational constant). Hence, we have a set of five equations here (cĀ = ħ = kBĀ = keĀ = G = 1), and so we can solve that to get the five Planck units, i.e. the Planck length unit, the Planck time unit, the Planck mass unit, the Planck energy unit, the Planck charge unit and, finally (oft forgotten), the Planck temperature unit. Of course, you should note that all mass and energy units areĀ directlyĀ related because of the mass-energy equivalence relationĀ E = mc2, which simplifies to E = m if cĀ is equated to 1. [I could also say something about the relation between temperature and (kinetic) energy, but I won’t, as it would only further confuse you.]

OK. Done! šŸ™‚

Addendum:Ā How to think about space and time?

If you read the argument on the Planck scale and constant carefully, then you’ll note that it does notĀ depend on the idea of an indivisible photon. However, it does depend on that Planck-Einstein relation being valid always and everywhere. Now, the Planck-Einstein relation is, in its essence, a fairly basic result from classical electromagnetic theory: itĀ incorporatesĀ quantum theory – remember: it’s the equation that allowed Planck to solve the black-body radiation problem, and so it’s why they call Planck the (reluctant) ‘Father of Quantum Theory’ – but it’sĀ notĀ quantum theory.

So the obvious question is: can we make this reflection somewhat more general, so we can think of theĀ electromagneticĀ force as an example only. In other words:Ā can we apply the thoughts above to any force and any movement really?

The truth is: I haven’t advanced enough in my little study to give the equations for the other forces. Of course, we could think of gravity, and I developed some thoughtsĀ on how gravity waves might look like, but nothing specific really. And then we have the shorter-range nuclear forces, of course: the strong force, and the weak force. The laws involved areĀ very different. The strong force involves colorĀ charges, and the way distances work is entirely different. So it would surely be some differentĀ analysis. However, the results should be the same. Let me offer some thoughts though:

  • We know that the relative strength of the nuclear force is much larger, because it pulls like charges (protons) together, despite the strong electromagnetic force that wants to push them apart! So the mentioned problem of trying to ‘pack’ some oscillation in some tiny little space should be worse with the strong force. And the strong force is there, obviously, at tiny little distances!
  • Even gravity should become important, because if we’ve got a lot of energy packed into some tiny space, its equivalent mass will ensure the gravitational forces also become important. In fact, that’s what the whole argument was all about!
  • There’s also all this talk about the fundamental forces becoming one at the Planck scale. I must, again, admit my knowledge is not advanced enough to explain how that would be possible, but I must assume that, if physicists are making such statements, the argument must be fairly robust.

So…Ā Whatever charge or whatever force we are talking about, we’ll be thinking of waves or oscillations—or simply movement, but it’s always a movement in a force field, and so there’s power and energy involved (energy is force times distance, and power is the time rate of change of energy). So, yes, we should expect the same issues in regard to scale. And so that’s what’s captured byĀ h.

As we’re talking the smallest things possible, I should also mention that there are also other inconsistencies in the electromagnetic theory, which should (also) have their parallel for other forces. For example, the idea of aĀ point chargeĀ is mathematically inconsistent, as I show in my post on fields and charges. Charge, any charge really, must occupyĀ someĀ space. It cannotĀ all be squeezed into one dimensionless point. So the reasoningĀ behind the Planck time and distance scale is surely valid.

In short, the whole argument about the Planck scale and those limits is veryĀ valid. However, does it imply our thinking about the Planck scale is actually relevant? I mean: it’s not because we canĀ imagineĀ how things might look like  – they may lookĀ like those tiny little black holes, for example – that these things actuallyĀ exist.Ā GUT or string theorists obviously think they are thinking about somethingĀ real. But, frankly, Feynman had a point when he said what he said about string theory, shortly before his untimely death in 1988: “I don’t like that they’re not calculating anything. I don’t like that they don’t check their ideas. I don’t like that for anything that disagrees with an experiment, they cook up an explanation—a fix-up to say, ‘Well, it still might be true.'”

It’s true that the so-called Standard Model doesĀ notĀ look very nice. It’s not like Maxwell’s equations. It’s complicated. It’s got various ‘sectors’: the electroweak sector, the QCD sector, the Higgs sector,… So ‘it looks like it’s got too much going on’, as a friend of mine said when he looked at a new design for mountainbike suspension. šŸ™‚ But, unlike mountainbike designs, there’s no real alternative for the Standard Model. So perhaps we should just accept it is what it is and, hence, in a way, accept Nature as we canĀ see it. So perhaps we should just continue to focus on what’s here, before we reach the Great Desert, rather than wasting time on trying to figure out how things might look like on the other side, especially because we’ll never be able to test our theories about ‘the other side.’

On the other hand, we can see where the Great Desert sort of startsĀ (somewhere near the 1032Ā Hz scale), and so it’s only natural to think it should alsoĀ stop somewhere. In fact, we knowĀ where it stops: it stops at the 1043Ā Hz scale, because everything beyond that doesn’t make sense. The question is: is there actually there? Like fundamental strings or whatever you want to call it. Perhaps we should just stop where the Great Desert begins. And what’s the Great Desert anyway? Perhaps it’s a desert indeed, and so then there is absolutelyĀ nothingĀ there. šŸ™‚

Hmm… There’s not all that much one can say about it. However, when looking at the history of physics, there’s one thing that’s really striking. Most of what physicists can think of,Ā in the sense that it made physical sense, turned out toĀ exist. Think of anti-matter, for instance. Paul Dirac thought it might exist, that it made sense to exist, and so everyone started looking for it, and Carl Anderson found in a few years later (in 1932). In fact, it had been observed before, but people just didn’t pay attention, so they didn’t want to see it, in a way. […] OK. I am exaggerating a bit, but you know what I mean. The 1930s are full of examples like that. There was a burst of scientific creativity, as the formalism of quantum physics was being developed, and the experimental confirmations of the theory just followed suit.

In the field of astronomy, or astrophysics I should say, it was the same with black holes.Ā No one could really imagine the existence of black holes until the 1960s or so: they were thought of a mathematical curiosity only, a logical possibility. However, the circumstantial evidence now is quite large and so… Well… It seems a lot of what we can think ofĀ actually has some existence somewhere. šŸ™‚

So… Who knows?Ā […] I surely don’t. And so IĀ need to get back toĀ the grind and work my way through the rest of Feynman’s LecturesĀ and the related math. However, this was a nice digression, and so I am grateful to my brother he initiated it. šŸ™‚

Some content on this page was disabled on June 17, 2020 as a result of a DMCA takedown notice from Michael A. Gottlieb, Rudolf Pfeiffer, and The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/

Self-inductance, mutual inductance, and the power and energy of inductors

As Feynman puts it, studying physics is not always about the ‘the great and esoteric heights’. In fact, you usuallyĀ have to come down from then fairly quickly – studying physics is similar to mountain climbing in that regard šŸ™‚ – and study ‘relatively low-level subjects’, such as electrical circuits, which is what we’ll do in this and the next post.

As I’ve introduced some key concepts in a previous post already, let me recapitulate the basics, which include the concept of theĀ electromotive force, which is basically theĀ voltage, i.e. the potential difference, that’s produced in a loop or coil of wire as the magnetic flux changes. I also talked about theĀ impedanceĀ in a AC circuit. Finally, we also discussed the powerĀ andĀ energiesĀ involved. Important results from this previous discussion include (but are not limited to):

  1. A constant speed AC generator will create an alternating current with the emf, i.e. the voltage, varying as V0Ā·sin(ωt).
  2. If we only have resistors as circuit elements, and the resistance in the circuit adds up to R, then the electric current in the circuit will be equal to I = ʐ/R = V/R = (V0/R)Ā·sin(ωt). So that’s Ohm’s Law, basically.
  3. The power that’s produced and consumed in an AC circuit is the product of the voltage and the current, so P = Ɛ·I = VĀ·I. We also showed this electrical power is equal to theĀ mechanicalĀ power dW/dt that makes the generator run.
  4. Finally, we explained the concept of impedance (denoted by Z) using Euler’s formula:Ā Z =Ā |Z|eiĪø, mentioning that, if other circuit elements than resistors are involved, such as inductors,Ā then it’s quite likely that the current signal will lag the voltage signal, with the phase factor Īø telling us by how much.

It’s now time to introduce those ‘other’ circuit elements. So let’s start with inductors here, and the concept ofĀ inductanceĀ itself. There’s a lot of stuff to them, and so let’s go over it point by point.

The concept of self-inductance

In its simplest form, an inductor is just a coil, but they come in all sizes and shapes. If you want to see how they might look like, justĀ googleĀ some images of micro-electronic inductors, and then, just to see the range of applications, some images of inductors used for large-scale industrial applications. If you do so, you’re likely to see images ofĀ transformersĀ too, because transformers work on the principle ofĀ mutualĀ inductance, and so they involveĀ twoĀ coils, i.e. two inductors.

Contrary to what you might expect, the concept of self-inductance (or inductance tout court) is quite simply: a changing current will cause a changingĀ magnetic field and, hence, some emf. Now, it turns out that the inducedĀ emfĀ is proportional to the change in current. So we’ve got another constant of proportionality here, so it’s like how we defined resistance, orĀ capacitance. So, in many ways, theĀ inductanceĀ is just another proportionality coefficient. If we denote it by L – the symbol is said to honor the Russian phyicistĀ Heinrich Lenz, whom you know from Lenz’ Law – then we define it as:

L = āˆ’Ę/(dI/dt)

The dI/dt factor is, obviously, the time rate of change of the current, and the negative sign indicates that the emfĀ opposesĀ the change in current, so it will tend to cause anĀ opposingĀ current. That’s why the emf involved is often referred to as a ‘back emf’. So that’sĀ Lenz’ LawĀ basically. As you might expect, the physicists came up with yet another derived unit, theĀ Henry, to honor yet another physicist, Joseph Henry, an American scientist who was a contemporary of Michael Faraday and independently discovered pretty much the same as Faraday: oneĀ henryĀ (H) equals one voltĀ·second per ampere: 1 H = 1 VĀ·s/A.

The concept of mutual inductance

FeynmanĀ introduces the topic of inductance with a two-coil set-up, as shown below, noting that a current in coil 1 will induce some emf in coil 2, which he denotes byĀ M12. Conversely, a current in coil 2 will induce some emf in coil 1, which he denoted by M21. M12Ā and M21Ā are also constants: they depend on the geometry of the coils, including the length of the solenoid (l), its surface area (S) and the number of loop turns of the coils (N1Ā and N2).

mutual inductance

The next step in the analysis is then to acknowledge that each coil should also produce a ‘back emf’ in itself, which we can denote by M11Ā and M22Ā respectively, but then these constants are, of course, equal to the self-inductanceĀ of the coils so, taking into account the convention in regard to the sign of the self-inductance, we write:

L1Ā = āˆ’ M11Ā and L1Ā =Ā āˆ’ M22

You will now wonder: what’s theĀ totalĀ emf in each coil, taking into account that we do not only have mutual inductance but also self-inductance? Frankly, when I was a kid, and my father tried to tell me one or two things about this, it confused me very much. I could not imagine what happened in one coil, let alone in twoĀ coils. I had this vision of a current producing some ‘back-current’, and then the ‘back-current’ producing ‘back-current’ again, and so I could not imagine how one could solve this problem. So the image in my head was very much like that baking powder box which Feynman evokes when talking about the method of imagesĀ to find the electric fields in situations with an easy geometry, so that’s the picture of a baking powder box which has on its label a picture of a baking powder box which has… Well… Etcetera. Unfortunately, my father didn’t push us to study math and, therefore, I knewĀ that one could solve such problems mathematically – we’re talking a converging series here – but I did notĀ know how, and that’s why I found it all veryĀ confusing.

Now I understand there’s one current only, and one potential difference only, and that the formulas doĀ notĀ involve some infinite series of terms. But… Well… I am not ashamed to say these problems are still testing the (limited) agility of my mind. The first thing to ‘get’ is that we’re talking a backĀ emf, and so that’s not a current but a potential difference. In fact, as I explained in my post on the electromotive force, the term ‘force’ in emf is actually misleading, and may lead to that same erroneous vision that I had as a kid: forces generating counter-forces, that generate counter-forces, that generate counter-forces, etcetera. It’s not like that: we have some current – oneĀ current – in a coil and we’ll have someĀ voltage – oneĀ voltage – across the coil. If the coil would be a resistorĀ instead of a coil, we’d find that the ratio of this voltage and the current would be some constant R = V/I. Now here we’re talking a coil indeed, so that’s aĀ differentĀ circuit element, and we find some other ratio, L =Ā āˆ’V/(dI/dt) = āˆ’Ę/(dI/dt). Why the minus sign? Well… As said, the inducedĀ emf will be such that it will tend to counter the current, and current flows fromĀ positiveĀ toĀ negativeĀ as per our convention.

But…Ā Yes?Ā So how does it work when we put this coil in some circuit, and how does the resistance of the inductor come into play? Relax. We’ve just been talking ideal circuit elementsĀ so far, and we’ve discussed only two: the resistor and the inductor. We’ll talk about voltage sources (or generators) and capacitors too, and then we’ll link all of these ideal circuit elements. In short, we’ll analyze some real-life electricalĀ circuitĀ soon, but first you need to understand the basics. Let me just note that an ideal inductor appears a zero-resistance conductor in a direct currentĀ (DC) circuit, so it’s a short-circuit really! Please try to mentallyĀ separate out those ‘ideal’ circuit components. Otherwise you’ll never be able to make sense of it all!

In fact, there’s a good reason why Feynman starts with explainingĀ mutual inductance before discussing a little circuit like the one below, which has an inductorĀ andĀ a voltage source. The two-coil situation above is effectively easier to understand, although it may not look like that at first. So let’s analyze that two-coil situation in more detail first. In other words, let me try to understand the situation that I didn’t understand as a kid. šŸ™‚

circuit with coil

Because of the law of superposition, we should add fluxes and changes in fluxes and, hence, we should also add the electromotive forces, i.e. the induced voltages. So, what we have here is that the total emf in coil 2 should be written as:

ʐ2Ā = M21Ā·(dI1/dt) +Ā M22Ā·(dI2/dt) =Ā M21Ā·(dI1/dt) – L2Ā·(dI2/dt)

What we’re saying here is that the emf, i.e. the voltage across the coil, will indeed depend on the change in current in the other coil, but also on the change in current of the coil itself. Likewise, the total emf in coil 1 should be written as:

ʐ1Ā = M12Ā·(dI2/dt) +Ā M11Ā·(dI1/dt) =Ā M12Ā·(dI2/dt) – L1Ā·(dI1/dt)

Of course, this does reduce to the simple L = āˆ’Ę/(dI/dt)Ā if there’s one coil only. But so you see where it comes from and, while we doĀ notĀ have some infinite series šŸ™‚ we do have a system of two equations here, and so let me say one or two things about it.

The first thing to note is that it is not so difficult to show thatĀ M21Ā is equal toĀ M12, so we can simplify and write that M21Ā = M12Ā = M. Now, while I said ‘not so difficult’, I didn’t mean it’s easy and, because I don’t want this post to become too long, I’ll refer you to Feynman for the proof of this M21Ā = M12Ā = M equation. It’s a general proof for any two coils or ‘circuits’ of arbitrary shape and it’s really worth the read. However, I have to move on.

The second thing to note is that this coefficientĀ M, which is referred to as theĀ mutual inductance now (so singular instead of plural) depends on the ‘circuit geometry’ indeed. For a simple solenoid, Feynman calculates it as

M = āˆ’(1/ε0c2)Ā·(N1Ā·N2)Ā·S/l,

withĀ lĀ the length of the solenoid, SĀ its surface area (S), and N1Ā and N2Ā the respective number of loop turns of the two coils. So, yes, only ‘geometry’ comes into play. [Note that’s quite obvious from the formula because a switch of the subscriptsĀ ofĀ N1Ā and N2Ā makes no difference, of course!]. Now, it’s interesting to note that M is the same for, let’s say, N1Ā = 100 and N2Ā = 10 and for N1Ā = 20 and N2Ā = 50. In fact, because you’re familiar with what transformers do, i.e. transforming voltages, you may think that’s counter-intuitive. It’s not. The thing with the number of coils does notĀ imply that ʐ1Ā and Ɛ2Ā remain the same. Our set of equations is Ɛ1Ā = MĀ·(dI2/dt) – L1Ā·(dI1/dt) and Ɛ2Ā = MĀ·(dI1/dt) – L2Ā·(dI2/dt), and so L1Ā and L2Ā clearly do vary asĀ N1Ā and N2Ā vary! So… Well… Yes. We’ve got a set of two equations with two independent variables (I1Ā and I2) and two dependent variables (ʐ1Ā and ʐ2). Of course, we could also phrase the problem the other way around: given two voltages, what are the currents? šŸ™‚

Of course, that makes us think of the power that goes in and out of a transformer. Indeed, you’ll remember that power is voltage times current. So what’s going on here in regard to that?

Well… There’s a thing with transformers, or with two-coil systems like this in general, that is referred to as coupling. The geometry of the situation will determineĀ how much flux from one coil is linked with the flux of the other coil. If most, or all of it, is linked, we say the two coils are ‘tightly coupled’ or, in the limit, that they are fully coupled. There’s a measure for that, and it’s called theĀ coefficient of coupling. Let’s first explore that concept of power once more.

Inductance, energy and electric power

It’s easy to see that we needĀ electric powerĀ to get some current going. Now, as we pointed out in our previous post, the power is equal to the voltage times the current. It’s also equal, of course, to the amount of work done per second, i.e. the time rate of change of the energy W, so we write:

dW/dt = Ɛ·I

Now, we defined the self-inductance as L = āˆ’Ę/(dI/dt) and, therefore, we know that Ɛ = āˆ’LĀ·(dI/dt), so we have:

dW/dt = āˆ’LĀ·IĀ·(dI/dt)

What is this? A differential equation? Yes and no. We’ve got not one but two functions of time here, W and I, and, while their derivatives with respect to time do appear in the equation, what we need to do is just integrate the two sides over time. We get: W =Ā āˆ’(1/2)Ā·LĀ·I2. Just check it by taking the time derivative of both sides. Of course, we can add any constant, to both sides in fact, but that’s just a matter of choosing some reference point. We’ll chose our constant to be zero, and also think about the energy that’sĀ storedĀ in the coil, i.e. U, which we define as:

U =Ā āˆ’W =Ā āˆ’(1/2)Ā·LĀ·I2Ā 

Huh?Ā What’s going on here? Well… It’s not an easy discussion, but let’s try to make sense of it. We have someĀ changingĀ current in the coil here but, obviously, some kind of inertia also: the coil itself opposes the change in current through the ‘back emf’. It requires energy, or power, to overcome the inertia. We may think of applying some voltage to offset the ‘back emf’, so we may effectively think of that little circuit with an inductorĀ and a voltage source. The voltage V we’d need to apply to offset the inertia would, obviously, be equal to the ‘back emf’, but with its sign reversed, so we have:

V = āˆ’Ā ĘĀ = LĀ·(dI/dt)

Now, it helps to think of what a current really is: it’s about electric charges that are moving at some velocity v because some force is being applied to them. As in any system, the power that’s being delivered is the dot product of the force and the velocity vectors (that ensures we only take the tangential component of the force into account), so if we have nĀ moving charges, the power that is being delivered to the circuit isĀ (FĀ·v)Ā·n. What is F? It’s obviously, qE, as the electric field is the force per unit charge, so E = F/q. But so we’re talking some circuitĀ here and we need to think ofĀ the power being delivered to some infinitesimal element ds in the coil, and so that’s (FĀ·v)Ā·nĀ·ds, which can be written as:Ā (FĀ·ds)Ā·nĀ·v. And then we integrate over the whole coil to find:power

Now, you may or may not remember that the emf (Ɛ) is actually defined as the line integral ∫ E·ds line, taken around the entire coil and, hence, noting that E = F/q, and that the current I is equal to I = q·n·v, we got our power equation. Indeed, the integrand or kernel of our integral becomes F·n·v·ds = q·E·n·v·ds = I·E·ds. Hence, we get our power formula indeed: P = V·I, with V the potential difference, i.e. the voltage across the coil.

I am getting too much into the weeds here. The point is: we’ve got a full and complete analog to the concept of inertia in mechanics here: instead of some forceĀ F causing someĀ massĀ m to change its velocity according to Newton’s Law, i.e. F = mĀ·a = mĀ·(dv/dt), we here have a potential difference V causing some current I to change according to the V = LĀ·(dI/dt) law.

This is very confusing but, remember, the same equations must have the same solutions! So, in an electric circuit, the inductance is really like what the mass is in a mechanics. Now, in mechanics, we’ll say that our mass has some momentum p = mĀ·v, and we’ll also say that its kinetic energy is equal to (1/2)mĀ·v2. We can do the same for our circuit: potential energy is continuously being converted into kinetic energy which, for our inductor, we write as U = (1/2)Ā·LĀ·I2.

Just think about by playing with one of the many online graphing tools. The graph below, for example, assumes the current builds up to some maximum. As it reaches its maximum, theĀ storedĀ energy will also max out. Now, you should not worry about the units here, or the scaleĀ of the graphs. The assumption is that I builds up from 0 to 1, and that L = 1, so that makes U what it is. Using a different constant for L, and/or different units for I, will change the scale of U too, but not its general shape, and that shape gives you the general idea.

powerThe example above obviously assumes some direct current, so it’s a DC circuit: the current builds up, but then stabilizes at some maximum that we can find by applying Ohm’s Law to the resistanceĀ of the circuit: I = V/R. Resistance? But we were talking an ideal inductor? We are. If there’s no other resistance in the circuit, we’ll have a short-circuit, so the assumption is that we doĀ have some resistance in the circuit and, therefore, we should also think of some energyĀ lossĀ to heat from the current in the resistance, but that’s not our worry here.

The illustration below is, perhaps, more interesting. Here we are, obviously, applying anĀ alternating current, and so the current goes in one and then in the other direction, so I > 0, and then I < 0, etcetera. We’re assuming some nice sinusoidal curve for the current here (i.e. the blue curve), and so we get what we get for U (i.e. the red curve): the energy goes up and down between zero and some maximum amplitude that’s determined by the maximum current.

power 2

So, yes, it is, after all, quite intuitive: building up a current does require energy from some external source, which is used to overcome the ‘back emf’ in the inductor, and that energy is stored in the inductor itself. [If you still wonder why it’s stored in the inductor, think about the other question: where else would it be stored?]Ā HowĀ is stored? Look at the graph and think: it’s stored as kinetic energy of the charges, obviously. That explains why the energy is zero when the current is zero, and why the energy maxes out when the current maxes out. So, yes, it all makes sense! šŸ™‚

Let’s now get back to that coupling constant.

The coupling constant

We can apply our reasoning to two coils. Indeed, we know that Ɛ1Ā = MĀ·(dI2/dt) – L1Ā·(dI1/dt) and Ɛ2Ā = MĀ·(dI1/dt) – L2Ā·(dI2/dt). So the power in the two-coils system is dW/dt = Ɛ1Ā·I1Ā + ʐ2Ā·I2, so we have:

dW/dt = MĀ·I1(dI2/dt) – L1Ā·I1Ā·(dI1/dt) + MĀ·I2Ā·(dI1/dt) – L2Ā·I2Ā·(dI2/dt)

= – L1Ā·I1Ā·(dI1/dt) – L2Ā·I2Ā·(dI2/dt) + MĀ·I1(dI2/dt)Ā·I2Ā·(dI1/dt)

Integrating both sides, and equating U with āˆ’W once more, yields:

U = (1/2)Ā·L1Ā·I12Ā + (1/2)Ā·L2Ā·I22Ā + MĀ·I1Ā·I2

[Again, you should just take the time derivative to verify this. If you don’t forget to apply the product rule for the MĀ·I1Ā·I2Ā term, you’ll see I am not writing too much nonsense here.] Now, there’s an interestingĀ algebraic transformation of this expression, and an equally interesting explanationĀ whyĀ we’d re-write the expression as we do. Let me copy it fromĀ Feynman so I’ll be using his fancier L and M symbols now. šŸ™‚

explanation coupling constant

So what? Well… Taking into account that inequality above, we can write the relation between M and the self-inductancesĀ L1Ā andĀ L2Ā using some constant k, which varies between 0 and 1 and which we’ll refer to as theĀ coupling constant:

formula coupling constant 2

We refer to k as the coupling constant, for rather obvious reasons: if it’s near zero, the mutual inductance will be very small, and if it’s near one, then the coils are said to be ‘tightly coupled’, and the ‘mutual flux linkage’ is then maximized. As you can imagine, there’s a whole body of literature out there relating this coupling constant to the behavior of transformers or other circuits where mutual inductance plays a role.

The formula for self-inductance

We gave the formula for theĀ mutualĀ inductance of two coils that are arranged as one solenoid on top of the other (cf. the illustration I started with):

M = āˆ’(1/ε0c2)Ā·(N1Ā·N2)Ā·S/l

It’s a very easy calculation, so let me quickly copy it from Feynman:

calculation solenoid

You’ll say: where is the M here? This is a formula for the emf! It is, but M is the constant of proportionality in front, remember? So there you go. šŸ™‚

Now, you would think that getting a formula for the self-inductance L of some solenoid would be equally straightforward. It turns out that that isĀ notĀ the case. Feynman needs two full pages and… Well… By now, you should now how ‘dense’ his writing really is: if it weren’t so dense, you’d be reading Feynman yourself, rather than my ‘explanations’ of him. šŸ™‚ So… Well.. If you want to see how it works, just click on the link hereĀ and scroll down to the last two pages of hisĀ exposé on self-inductance. I’ll limit myself to just jotting down the formula he does obtain when he’s through the whole argument:

solenoid formula

See why he uses a fancier L than ‘my’ L? ‘His’ L is the length of the solenoid. šŸ™‚ And, yes, r is the radius of the coil and n the number of turnsĀ per unit lengthĀ in the winding. Also note this formula is valid only if L >> R, so the effects at theĀ endĀ of the solenoid can be neglected. OK. Done. šŸ™‚

Well… That’s it for today! I am sorry to say but the next post promises to be as boring as this one because… Well… It’s on electric circuits again. 😦

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 20, 2020 as a result of a DMCA takedown notice from Michael A. Gottlieb, Rudolf Pfeiffer, and The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/

Reconciling the wave-particle duality in electromagnetism

As I talked about Feynman’s equation for electromagnetic radiation in my previous post, I thought I should add a few remarks on wave-particle duality, but then I didn’t do it there, because my post would have become way too long. So let me add those remarks here. In fact, I’ve written about this before, and so I’ll just mention the basic ideas without going too much in detail.Ā Let me first jot down the formula once again, as well as illustrate the geometry of the situation:

formual

geometry

The gist of the matter is that light, in classical theory, is a traveling electromagnetic field caused by an accelerating electric charge and that, because light travels at speed c, it’s the acceleration at the retarded time t – r/c, i.e. a‘ = a(t – r/c), that enters the formula. You’ve also seen the diagrams that accompany this formula:

EM 1 EM 2

The two diagrams above show that the curve of the electric fieldĀ in spaceĀ is a ā€œreversedā€ plot of the acceleration as a function of time. As I mentioned before, that’s quite obvious from the mathematical behavior of a function with argument like the argument above, i.e. a function F(t – r/c). When we write t – r/c, we basically measureĀ distance units in seconds, instead of in meter. So we basically useĀ cĀ as theĀ scale for both time as well as distance. I explained that in a previous post, so please have a look there if you’d want so see how that works.

So it’s pretty straightforward, really. However, having said that, when I see a diagram like the one above, so all of these diagrams plotting anĀ E or BĀ waveĀ in space, I can’t help thinking it’s somewhat misleading: after all, we’re talking something traveling at the speed of light here and, therefore, its length – inĀ ourĀ frame of reference – should be zero. And it is, obviously. Electromagnetic radiation comes packed in point-like, dimensionless photons: the length of something that travels at the speed of lightĀ mustĀ be zero.

Now, I don’t claim to know what’s going onĀ exactly, but my thinking on it may not be far off the mark. We know that light is emitted and absorbed by atoms, as electrons go from one energy level to another, and the energy of theĀ photonsĀ of lightĀ corresponds to the differenceĀ between those energy levels (i.e. a few electron-volt only, typically: it’s given by the E = h·ν relation). Therefore, we can look at a photon as aĀ transient electromagnetic wave. It’s a very short pulse: the decay time for one such pulse ofĀ sodium light, i.e. one photon of sodium light,Ā isĀ 3.2Ɨ10–8Ā seconds. However, taking into account the frequency of sodium light (500 THz), that still makes for some 16 million oscillations, and a wave-train with a length of almost 10 meter. [Yes. Quite incredible, isn’t it?]Ā SoĀ the photon could look like the transient wave I depicted below, except… Well… This wavetrain is traveling at the speed of light and, hence, we will not see it as a ten-meter long wave-train. Why not? Well…Ā Because of the relativistic length contraction, it will effectively appear as a point-like particle to us.

Photon wave

So relativistic length contraction is whyĀ the wave and particle duality can be easily reconciled in electromagnetism: we can think of light as an irregular beam of point-like photons indeed, as one atomic oscillator after the other releases a photon, in no particularly organized way. So we can think of photons asĀ transient wave-trains, but we should remind ourselves that they are traveling at the speed of light, so they’ll look point-like to us.

Is such view consistent with theĀ results of the famous – of should I say infamous? – double-slit experiment. Well… Maybe. As I mentioned in one of my posts, it is rather remarkable that is actually hard to find actual double-slitĀ experiments that use actualĀ detectors near the slits, and even harder to find such experiments involving photons! Indeed, experiments involving detectors near the slits are usually experiments with ‘real’ particles, such as electrons, for example. Now, a lot of advances have been made in the set-up of these experiments over the past five years, and one of these experiments isĀ a 2010 experiment of an Italian teamĀ which suggests that it’s the interaction between the detector and the electron waveĀ that may cause the interference pattern to disappear. Now thatĀ throws some doubts on the traditional explanation of the results of the double-slit experiment.

The idea is shown below. The electron is depicted as an incoming plane wave which effectively breaks up as it goes through the slits. The slit on the left has no ā€˜filter’ (which you may think of as a detector) and, hence, the plane wave goes through as a cylindrical wave. The slit on the right-hand side isĀ covered by a ā€˜filter’ made of several layers of ā€˜low atomic number material’, so the electron goes through but, at the same time, the barrier creates a spherical wave as it goes through. The researchers note that ā€œthe spherical and cylindrical wave do not have any phase correlation, and so even if an electron passed through both slits, the two different waves that come out cannot create an interference pattern on the wall behind them.ā€ [I hope I don’t have to remind you that, while being represented as ā€˜real’ waves here, the ā€˜waves’ are, obviously, complex-valued psi functions.]

double-slit experiment

In fact, to be precise, the experimenters note that there still was an interference effect if the filter was thin enough. Let me quote the reason for that: ā€œThe thicker the filter, the greater the probability for inelastic scattering. When the electron suffers inelastic scattering, it is localized. This means that its wavefunction collapses and, after the measurement act, it propagates roughly as a spherical wave from the region of interaction, with no phase relation at all with other elastically or inelastically scattered electrons. If the filter is made thick enough, the interference effects cancels out almost completely.ā€

This does notĀ solve the ‘mystery’ of the double-slit experiment, but it throws doubt on how it’s usually being explained. The mystery in such experiments is that, when we put detectors, it isĀ eitherĀ theĀ detector at AĀ or the detector at B that goes off. They should never go off togetherā€”ā€at half strength, perhapsā€, as Feynman puts it. But so there areĀ doubts here now. Perhaps the electron doesĀ go through both slits at the same time! And soĀ that’s why I used italics when writing ā€œeven if an electron passed through both slitsā€: the electron, or the photon in a similar set-up, is not supposed to do that according to the traditional explanation of the results of the double-slit experiment! It’s one or the other, and the wavefunction collapsesĀ or reduces as it goes through.Ā 

However,Ā that’s where these so-called ā€˜weak measurement’ experiments now come in, like this 2010 experiment: it does not prove but indicates that interaction does not have to be that way. They strongly suggest that it is notĀ all or nothing, that our observations should not necessarily destroy the wavefunction. So, who knows, perhaps we will be able, one day, to show that the wavefunction does go through both slits, as it should (otherwise the interference pattern cannot be explained), and then we will have resolved the paradox.

I am pretty sure that, when that’s done, physicists will also be able to relate the image of a photon as a transient electromagnetic wave (cf. the diagram above), being emitted by an atomic oscillator for a few nanoseconds only (we gave the example for sodium light, for which the decay time was 3.2Ɨ10–8Ā seconds) with the image of a photon as a particle that can be represented by a complex-valued probability amplitude functionĀ (cf. the diagram below). I look forward to that day. I think it will come soon.

Photon wave

Here I should add two remarks. First, a lot hasĀ been said about the so-called indivisibility of a photon, but inelastic scattering implies that photons are notĀ monolithic: the photon loses energy to the electron and, hence, its wavelength changes. Now, you’ll say: the scattered photon is not the same photon as the incident photon, and you’re right. But… Well. Think about it. It does say something about the presumed oneness of a photon.

I15-72-Compton1

The other remark is on the mathematics of interference. Photons are bosons and, therefore, we have to add their amplitudes to get the interference effect. So you may try to think of an amplitude function, like ĪØ = (1/√2Ļ€)Ā·eiĪøĀ or whatever, and think it’s just a matter of ‘splitting’ this function before it enters the two slits and then ‘putting it back together’, so to say, after our photon has gone through the slits. [For the detailed math of interference in quantum mechanics, see my page on essentials.] Ā Well… No. It’s not that simple. The illustration with that plane wave entering the slits, and the cylindrical and/or spherical wave coming out, makes it obvious that something happens to our wave as it goes through the slit. As I said a couple of times already, the two-slit experiment is interesting, but the interference phenomenon – or diffraction as it’s called – involving one slit only is at least as interesting. So… Well… The analysis isĀ notĀ that simple. Not at all, really. šŸ™‚

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/

The LiĆ©nard–Wiechert potentials and the solution for Maxwell’s equations

In my post on gauges and gauge transformations in electromagnetics, I mentioned the full and complete solution for Maxwell’s equations, using the electric and magnetic (vector) potential Φ and A. Feynman frames it nicely, so I should print it and put it on the kitchen door, so I can look at it everyday. šŸ™‚

frame

I should print the wave equation we derived in our previous post too. Hmm… Stupid question, perhaps, but why is there no wave equation above? I mean: in the previous post, we said the wave equation was the solution for Maxwell’s equation, didn’t we?Ā The answer is simple, of course: the wave equation is a solution for waves originating from some source and traveling through free space, so that’s a special case. Here we have everything. Those integrals ‘sweep’ all over space, and so that’sĀ realĀ space, which is full of moving charges and so there’s waves everywhere. So the solution above is far more general and captures it all: it’s the potential at every point in space, and at every point in time, taking into account whatever else is there, moving or not moving. In fact, it isĀ the general solutionĀ of Maxwell’s equations.

How do we find it? Well… I could copy Feynman’s 21stĀ LectureĀ but I won’t do that. The solution is based on the formula for Φ and A for a small blob of charge, and then the formulas above just integrate over all of space. That solution for a small blob of charge, i.e. a point charge really, was first deduced in 1898, by a French engineer: Alfred-Marie LiĆ©nard. However, his equations did not get much attention, apparently, because a German physicist, Emil Johann Wiechert, worked on the same thing and found the very same equations just two years later. That’s why they are referred to as theĀ LiĆ©nard-Wiechert potentials, so they both get credit for it, even if both of them worked it out independently. These are the equations:

electric potential

magnetic potential

Now, you may wonder why I am mentioning them, and you may also wonder how we get those integrals above, i.e. our general solution for Maxwell’s equations, from them. You can find the answer to your second question in Feynman’s 21st Lecture. šŸ™‚ As for the first question, I mention them because one can derive two other formulas for E and B from them. It’s the formulas that Feynman uses in his firstĀ Volume, when studying light:Ā E

B

Now you’ll probably wonder how we can get these twoĀ equations from theĀ LiĆ©nard-Wiechert potentials. They don’t look very similar, do they? No, they don’t. Frankly, I would like to give you the same answer as above, i.e. check it in Feynman’s 21st Lecture, but the truth is that the derivation is so long and tedious that even Feynman says one needs “a lot of paper and a lot of time” for that. So… Well… I’d suggest we just use all of those formulas and not worry too much about where they come from. If we can agree on that, we’re actually sort of finished with electromagnetism. All the chapters that follow Feynman’s 21stĀ LectureĀ are applicationsĀ indeed, so they do not add all that much to the coreĀ of the classicalĀ theory of electromagnetism.

So why did I write this post? Well… I am not sure. I guess I just wanted to sum things up for myself, so I can print it all out and put it on the kitchen door indeed. šŸ™‚ Oh, and now that I think of it, I should add one more formula, and that’s the formula for sphericalĀ waves (as opposed to theĀ planeĀ waves we discussed in my previous post). It’s a very simple formula, and entirely what you’d expect to see:

spherical wave

The S function is the source function, and you can see that the formula is a Coulomb-like potential, but with the retarded argument. You’ll wonder: what is ψ? Is it E or B or what? Well… You can just substitute: ψ can be anything. Indeed, Feynman gives a very general solution for any type ofĀ spherical wave here. šŸ™‚

So… That’s it, folks. That’s all there is to it. I hope you enjoyed it. šŸ™‚

Addendum: Feynman’s equation for electromagnetic radiation

I talked about Feynman’s formula for electromagnetic radiation before, but it’s probably good to quickly re-explain it here. Note that it talks about the electric field only, as the magnetic field is so tiny and, in any case, if we have E then we can find B. So the formula is:

E

The geometry of the situation is depicted below. We have some charge q that, we assume, is moving through space, and so it creates some field E at point P. TheĀ er‘Ā vector is the unit vector from P to Q, so it points at the charge. Well… It points to where the charge was at the time just a little while ago, i.e. at the timeĀ t – r‘/c. Why? Well… We don’t know where q is right now, because the field needs some time travel, we don’t know q right now, i.e. q at time t. It might be anywhere. Perhaps it followed some weird trajectory during the time r‘/c, like the trajectory below.

radiation formula

So our er‘Ā vector moves as the charge moves, and so it will also have velocity and, likely, some acceleration, but what we measure for its velocity and acceleration, i.e. the d(er)/dt and d2(er)/dt2Ā in that Feynman equation, is also theĀ retardedĀ velocity and theĀ retardedĀ acceleration. But look at the terms in the equation. The first two terms have a 1/r’2Ā in them, so these two effects diminish with the square of the distance. The first term is just Coulomb’s Law (note that the minus sign in front takes care of the fact that like chargesĀ repelĀ and so the E vector will point in the other way). Well… It is and it isn’t, because of the retarded time argument, of course. And so we have the second term, which sort of compensates for that. Indeed, the d(er)/dt is the time rate of change of erĀ and, hence, if r‘/c = Ī”t, then (r‘/c)Ā·d(er)/dt is a first-order approximation of Ī”er.

As Feynman puts it: “The second term is as though nature were trying to allow for the fact that the Coulomb effect is retarded, if we might put it very crudely. It suggests that we should calculate the delayed Coulomb field but add a correction to it, which is its rate of change times the time delay that we use. Nature seems to be attempting to guess what the field at the present time is going to be, by taking the rate of change and multiplying by the time that is delayed.”Ā In short, the first two terms can be written as E = āˆ’(q/4πε0)/r2Ā·[erĀ +Ā Ī”er] and, hence, it’s a sort of modified Coulomb Law that sort of tries to guess what the electrostatic field at P should be based on (a) what it is right now, and (b) howĀ q’s direction and velocity, as measured now, would change it.

Now, the third term has a 1/c2Ā factor in front but, unlike the other two terms, this effect doesĀ notĀ fall off with distance. So the formula belowĀ fully describes electromagnetic radiation, indeed, because it’sĀ the only important term when we get ‘far enough away’, with ‘far enough’ meaning that the parts that go as the square of the distance have fallen off so much that they’re no longer significant.

radiation formula 2Of course, you’re smart, and so you’ll immediately note that, as r increases, that unit vector keeps wiggling but that effect will also diminish. You’re right. It does, but in a fairly complicated way.Ā The acceleration ofĀ erĀ has two components indeed. One is the transverse or tangential piece, because the end of erĀ goes up and down, and the other is a radial piece because it stays on a sphere and so it changes direction. The radial piece is the smallest bit, and actually also varies as the inverse square ofĀ rĀ when rĀ is fairly large. The tangential piece, however, varies only inversely as the distance, so as 1/r. So, yes, the wigglings of erĀ look smaller and smaller, inversely as the distance, but the tangential piece is and remains significant, because it does not vary as 1/r2Ā but as 1/r only.Ā  That’s why you’ll usually see the law of radiation written in an even simpler way:

final law of radiation

This law reduces the whole effect to the component of the acceleration that is perpendicular to the line of sightĀ only. It assumes the distance is huge as compared to the distance over whichĀ the charge is moving and, therefore, that r‘ and r can be equated for all practical purposes. It also notes that the tangential piece is all that matters, and so it equatesĀ d2(er)/dt2Ā with ax/r. The whole thing is probably best illustrated as below: we have aĀ generatorĀ driving charges up and down in G – so it’s an antennaĀ really – and so we’ll measure a strong signal when putting the radiation detector D in position 1, but we’ll measure nothing in position 3. [The detector is, of course, another antenna, but with an amplifier for the signal.] But so here I am starting to talk about electromagnetic radiation once more, which wasĀ notĀ what I wanted to do here, if only because Feynman does a much better job at that than I could ever do. šŸ™‚radiator

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 17, 2020 as a result of a DMCA takedown notice from Michael A. Gottlieb, Rudolf Pfeiffer, and The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 20, 2020 as a result of a DMCA takedown notice from Michael A. Gottlieb, Rudolf Pfeiffer, and The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/

Traveling fields: the wave equation and its solutions

Pre-script (dated 26 June 2020): Our ideas have evolved into a full-blown realistic (or classical) interpretation of all things quantum-mechanical. In addition, I note the dark force has amused himself by removing some material. So no use to read this. Read my recent papers instead. šŸ™‚

Original post:

We’ve climbed a big mountain over the past few weeks, post by post, šŸ™‚ slowly gaining height, and carefully checking out the various routes to the top. But we are there now: we finally fully understand how Maxwell’s equations actuallyĀ work. Let me jot them down once more:

Maxwell's equations

As for how real or unreal the E and B fields are, I gave you Feynman’s answer to it, so… Well… I can’t add to that. I should just note, or remind you, that we have a fully equivalent description of it all in terms of the electric and magnetic (vector) potential Φ and A, and so we can ask the same question about Φ and A. They explain realĀ stuff, so they’reĀ real in that sense. That’s what Feynman’s answer amounts to, and I am happy with it. šŸ™‚

What I want to do here is show how we can get from those equations to some kind of wave equation: an equation that describes how a field actuallyĀ travelsĀ through space. So… Well…Ā Let’s first look at that very particular wave function we used in the previous post to prove that electromagnetic waves propagate with speedĀ c, i.e. the speed of light. The fields were very simple: the electric field hadĀ a y-component only, and the magnetic field a z-component only. Their magnitudes, i.e. their magnitude where the field had reached, as it fills the space traveling outwards,Ā were given in terms of J, i.e. the surface current density going in the positive y-direction, and the geometry of the situation is illustrated below.

equation

sheet of charge

The fields were, obviously, zero where the fields had not reached as they were traveling outwards. And, yes, I know that sounds stupid. But… Well… It’s just to make clear what we’re looking at here. šŸ™‚

We also showed how the wave would look like if we would turn off itsĀ First CauseĀ after some time T, so if the moving sheet of charge would no longer move after time T. We’d have the followingĀ pulse traveling through space, a rectangular shapeĀ really:

wavefrontWe can imagine more complicated shapes for the pulse, like the shape shown below. J goes from one unit to two units at time t =Ā t1Ā and then to zero at t =Ā t2. Now, the illustration on the right shows the electric field as a function of x at the time t shown by the arrow. We’ve seen this before when discussing waves: if the speed of travel of the wave is equal toĀ c, then x is equal to x = cĀ·t, and the pattern is as shown below indeed: it mirrors what happenedĀ at the source x/c secondsĀ ago. So we write:

equation 2

12

This idea of using theĀ retardedĀ time t’ = t āˆ’ x/c in the argument of a wave function f – or, what amounts to the same, using x āˆ’ c/t – is key to understanding wave functions. I’ve explained this in veryĀ simple language in a post for my kids and, if you don’t get this, I recommend you check it out. What we’re doing, basically, isĀ converting something expressed in time units intoĀ something expressed in distance units, or vice versa, using theĀ velocityĀ of the wave as the scale factor, so time and distance are bothĀ expressed in the same unit, which may be seconds, or meter.

To see how it works, suppose we addĀ some timeĀ Ī”t to the argument of our wave function f, so we’re looking at f[xāˆ’c(t+Ī”t)] now, instead of f(xāˆ’ct). Now, f[xāˆ’c(t+Ī”t)]Ā = f(xāˆ’ctāˆ’cĪ”t), so we’ll get a different valueĀ for our function—obviously! But it’s easy to see that we canĀ restore our wave function FĀ to its former value by alsoĀ adding some distanceĀ Ī”x =Ā cĪ”t to the argument. Indeed, if we do so, we get f[x+Ī”xāˆ’c(t+Ī”t)] = f(x+cĪ”t–ctāˆ’cĪ”t) = f(x–ct). You’ll say: t āˆ’ x/cĀ is not the same asĀ x–ct. It is and it isn’t: any function of x–ct is also a function of t āˆ’ x/c, because we can write:

capture

Here, I need to add something about the directionĀ of travel. The pulse above travel in the positive x-direction, so that’s why we haveĀ x minusĀ ct in the argument. For a wave traveling in the negative x-direction, we’ll have a wave function y = F(x+ct).Ā In any case, I can’t dwell on this, so let me move on.

Now, Maxwell’s equations in free or empty space, where are thereĀ no charges nor currentsĀ to interact with, reduce to:

Maxwell in free space

Now, how can we relate this set of complicated equations to a simple wave function? Let’s do the exercise for our simple EyĀ and BzĀ wave. Let’s start by writing out the first equation, i.e.Ā āˆ‡Ā·E = 0, so we get:

f1

Now, our wave does notĀ vary in the y and z direction, so none of the components, including EyĀ and EzĀ depend on y or z. It only varies in theĀ x-direction, soĀ āˆ‚Ey/āˆ‚y and āˆ‚Ez/āˆ‚z are zero. Note that the cross-derivatives āˆ‚Ey/āˆ‚z and āˆ‚Ez/āˆ‚y are also zero: we’re talking a plane wave here, the field varies only with x. However, because āˆ‡Ā·E = 0,Ā āˆ‚Ex/āˆ‚x must be zero and, hence, ExĀ must be zero.

Huh?Ā What?Ā How is that possible? You just said that our field doesĀ vary in the x-direction! And now you’re saying it doesn’t it? Read carefully. I know it’s complicated business, but it all makes sense. Look at the function: we’re talking Ey, notĀ Ex. EyĀ does vary as a function of x, but our field does not have an x-component, so ExĀ = 0. We have no cross-derivative āˆ‚Ey/āˆ‚x in the divergence of E (i.e. inĀ āˆ‡Ā·E = 0).

Huh?Ā What?Ā Let me put it differently.Ā E has three components: Ex,Ā EyĀ andĀ Ez, and we have three space coordinates: x, y and z, so we have nine cross-derivatives. What I am saying is that allĀ derivatives with respect to y and z are zero. That still leaves us with threeĀ derivatives:Ā āˆ‚Ex/āˆ‚x, āˆ‚Ey/āˆ‚x, andĀ āˆ‚Ey/āˆ‚x. So… Because all derivatives in respect to y and z are zero, and because of the āˆ‡Ā·E = 0 equation, we knowĀ thatĀ āˆ‚Ex/āˆ‚x must be zero.Ā So, to make a long story short, I did notĀ say anything aboutĀ āˆ‚Ey/āˆ‚x or āˆ‚Ez/āˆ‚x. These may still be whatever they want to be, and they may vary in more or in less complicated ways. I’ll give an example of that in a moment.

Having said that, I do agree that I was a bit quick in writing that, because āˆ‚Ex/āˆ‚x = 0, ExĀ must be zero too. Looking at the math only,Ā ExĀ is not necessarilyĀ zero: it might be some non-zero constant. So… Yes. That’s a mathematical possibility. The static field from some charged condenser plate would be an example of a constantĀ ExĀ field. However, the point is that we’re notĀ looking at such static fields here: we’re talking dynamicsĀ here, and we’re looking at a particular type of wave: we’re talking a so-called planeĀ wave. Now, the wave front of a plane wave is… Well… A plane. šŸ™‚ SoĀ ExĀ is zero indeed. It’s a general result for planeĀ waves: the electric field of a planeĀ wave will always be at right angles to the direction of propagation.

Hmm… I can feel your skepticism here. You’ll say I am arbitrarily restricting the field of analysis… Well… Yes. For the moment. It’s not a reasonableĀ restriction though. As I mentioned above, the field of a plane wave may still vary in both the y- and z-directions, as shown in the illustration below (for which the credit goes to Wikipedia), which visualizes the electric field of circularly polarized light. In any case, don’t worry too much about. Let’s get back to the analysis. Just note we’re talking plane waves here. We’ll talk about non-plane waves i.e.Ā incoherentĀ light waves later. šŸ™‚

circular polarization

So we have plane waves and, therefore, a so-called transverse E field which we can resolve in two components: EyĀ and Ez. However, we wanted to study a very simply EyĀ field only. Why? Remember the objective of this lesson: it’s just to show how we go from Maxwell’s equations to the wave function, and so let’s keep the analysis simple as we can for now: we can make it more general later.Ā In fact, ifĀ we do the analysis now for non-zero EyĀ and zero Ez, we can do a similar analysis forĀ non-zero EzĀ and zero Ey, and the general solution is going to be some superposition of two such fields, so we’ll have a non-zeroĀ EyĀ andĀ Ez. Capito? šŸ™‚ So let me write out Maxwell’s second equation, and use the results we got above, so I’llĀ incorporate the zero values for the derivatives with respect to y and z, and also the assumption that EzĀ is zero. So we get:

f3[By the way: note that, out of the nine derivatives, the curl involves only the (six) cross-derivatives. That’s linked to the neat separation between the curl and the divergence operator. Math is great! :-)]

Now, because of the flux rule (āˆ‡Ć—EĀ = ā€“āˆ‚B/āˆ‚t), we can (and should) equate the three components of āˆ‡Ć—E above with the three components of ā€“āˆ‚B/āˆ‚t, so we get:

f4

[In case you wonder what it is that I am trying to do, patience, please! We’ll get where we want to get. Just hang in there and read on.] Now, āˆ‚Bx/āˆ‚t = 0 and āˆ‚By/āˆ‚t = 0 do not necessarily imply that BxĀ andĀ ByĀ are zero: there might be some magnets and, hence, we may have some constant static field. However, that’s a matter of choosing a reference point or, more simply, assuming that empty space is effectively empty, and so we don’t have magnets lying around and so we assume that BxĀ andĀ ByĀ are effectively zero. [Again, we can always throw more stuff in when our analysis is finished, but let’s keep it simple and stupid right now, especially because the BxĀ = ByĀ = 0 is entirely in line with the ExĀ = EzĀ = 0 assumption.]

The equations above tell us what we know already: the E and B fields are at right angles to each other. However, note, once again, that this is a more general result for all planeĀ electromagneticĀ waves, so it’s not only that very special caterpillar or butterfly field that we’re looking at it. [If you didn’t read my previous post, you won’t get the pun, but don’t worry about it. You need to understand the equations, not the silly jokes.]

OK. We’re almost there. Now we need Maxwell’s last equation. When we write it out, we get the following monstrously lookingĀ set of equations:

f5

However, because of all of the equations involving zeroes above šŸ™‚ only āˆ‚Bz/āˆ‚x is notĀ equal to zero, so the whole set reduced to only simple equation only:

f6

Simplifying assumptions are great, aren’t they? šŸ™‚ Having said that, it’s easy to be confused. You should watch out for the denominators: a āˆ‚x and a āˆ‚t are two veryĀ different things. So we have two equations now involving first-orderĀ derivatives:

  1. āˆ‚Bz/āˆ‚t = āˆ’āˆ‚Ey/āˆ‚x
  2. āˆ’c2āˆ‚Bz/āˆ‚x = āˆ’āˆ‚Ey/āˆ‚t

So what? Patience,Ā please!Ā šŸ™‚ Let’s differentiate the first equation with respect to x and the second with respect to t. Why? Because… Well… You’ll see. Don’t complain. It’s simple. Just do it. We get:

  1. āˆ‚[āˆ‚Bz/āˆ‚t]/āˆ‚x = āˆ’āˆ‚2Ey/āˆ‚x2
  2. āˆ‚[āˆ’c2āˆ‚Bz/āˆ‚x]/āˆ‚t = āˆ’āˆ‚2Ey/āˆ‚x2

So we can equate the left-hand sides of our two equations now, and what we get is a differential equation of the second order that we’ve encountered already, when we were studying wave equations. In fact, itĀ isĀ the wave equation for one-dimensional waves:

f7In case you want to double-check, I did a few posts on this, but, if you don’t get this, well… I am sorry. You’ll need to do some homework. More in particular, you’ll need to do some homework on differential equations. The equation above is basically some constraintĀ on the functional form of Ey. More in general, if we see an equation like:

f8

then the function ψ(x, t) must be some function

solution

So any function ψ like that will work. You can check it out by doing the necessary derivatives and plug them into the wave equation. [In case you wonder how you should go about this, Feynman actually does it for you in his Lecture on this topic, so you may want to check it there.]

In fact, the functionsĀ f(x āˆ’ c/t) andĀ g(x + c/t) themselves will also work as possible solutions. So we can drop one or the other, which amounts to saying that our ‘shape’ has to travel in some direction, rather than in both at the same time.Ā šŸ™‚ Indeed, from all of my explanations above, you know whatĀ f(x āˆ’ c/t) represents: it’s a wave that travels in the positive x-direction. Now, it may be periodic, but it doesn’tĀ haveĀ to be periodic. TheĀ f(x āˆ’ c/t) function could representĀ any constant ‘shape’Ā that’s traveling in the positive x-direction at speed c. Likewise, theĀ g(x + c/t) function could representĀ any constant ‘shape’Ā that’s traveling in the negativeĀ x-direction at speed c. As for super-imposing both…

Well… I suggest you check that post I wrote for my son, Vincent. It’s on the math of waves, but it doesn’t have derivatives and/or differential equations. It just explains how superimposition and all that works. It’s not very abstract, as it revolves around a vibrating guitar string. So, if you have trouble with all of the above, you may want to read that first. šŸ™‚ The bottom line is that we can get anyĀ wavefunction we want by superimposing simple sinusoidals that are traveling in one or the other direction, and so that’s what’s the more general solution really says. Full stop. So that’s what’s we’re doing really: we add very simple waves to get very more complicated waveforms. šŸ™‚

Now, I could leave it at this, but then it’s veryĀ easy to just go one step further, and that is to assume that EzĀ and, therefore, ByĀ are not zero. It’s just a matter of super-imposing solutions. Let me just give you the general solution. Just look at it for a while. If you understood all that I’ve said above, 20 seconds or so should be sufficient to say: “Yes, that makes sense. That’s the solution in two dimensions.” At least, I hope so! šŸ™‚

General solution two dimensions

OK. I should really stop now. But… Well… Now that we’ve got a general solution for all plane waves, why not be even bolder and think about what we could possibly say aboutĀ three-dimensionalĀ waves? So then ExĀ and, therefore, BxĀ would not necessarily be zero either. After all, light can behave that way. In fact, light is likely to beĀ non-polarizedĀ and, hence, ExĀ and, therefore, BxĀ are most probably not equal toĀ zero!

Now, you may think the analysis is going to be terribly complicated. And you’re right. It would beĀ if we’d stick to our analysis in terms of x, y and z coordinates. However, it turns out that the analysis in terms of vector equations is actually quite straightforward. I’ll just copy the Master here, so you can see His Greatness. šŸ™‚

waves in three dimensions

But what solution does an equation like (20.27) have? We can appreciate it’s actually three equations, i.e. one for each component, and so… Well… Hmm… What can we say about that? I’ll quote the Master on this too:

“How shall we find the general wave solution? The answer is that all the solutions of the three-dimensional wave equation can be represented as a superposition of the one-dimensional solutions we have already found. We obtained the equation for waves which move in the x-direction by supposing that the field did not depend on yĀ andĀ z. Obviously, there are other solutions in which the fields do not depend on xĀ andĀ z, representing waves going in the y-direction. Then there are solutions which do not depend on xĀ andĀ y, representing waves travelling in the z-direction. Or in general, since we have written our equations in vector form, the three-dimensional wave equation can have solutions which are plane waves moving in any direction at all. Again, since the equations are linear, we may have simultaneously as many plane waves as we wish, travelling in as many different directions. Thus the most general solution of the three-dimensional wave equation is a superposition of all sorts of plane waves moving in all sorts of directions.”

It’s the same thing once more:Ā we add very simple waves to get very more complicated waveforms. šŸ™‚

You must have fallen asleep by now or, else, be watching something else. Feynman must have felt the same.Ā After explaining all of the nitty-gritty above, Feynman wakes up his students. He does so by appealing to their imagination:

“Try to imagine what the electric and magnetic fields look like at present in the space in this lecture room. First of all, there is a steady magnetic field; it comes from the currents in the interior of the earth—that is, the earth’s steady magnetic field. Then there are some irregular, nearly static electric fields produced perhaps by electric charges generated by friction as various people move about in their chairs and rub their coat sleeves against the chair arms. Then there are other magnetic fields produced by oscillating currents in the electrical wiring—fields which vary at a frequency of 6060Ā cycles per second, in synchronism with the generator at Boulder Dam. But more interesting are the electric and magnetic fields varying at much higher frequencies. For instance, as light travels from window to floor and wall to wall, there are little wiggles of the electric and magnetic fields moving along at 186,000Ā miles per second. Then there are also infrared waves travelling from the warm foreheads to the cold blackboard. And we have forgotten the ultraviolet light, the x-rays, and the radiowaves travelling through the room.

Flying across the room are electromagnetic waves which carry music of a jazz band. There are waves modulated by a series of impulses representing pictures of events going on in other parts of the world, or of imaginary aspirins dissolving in imaginary stomachs. To demonstrate the reality of these waves it is only necessary to turn on electronic equipment that converts these waves into pictures and sounds.

If we go into further detail to analyze even the smallest wiggles, there are tiny electromagnetic waves that have come into the room from enormous distances. There are now tiny oscillations of the electric field, whose crests are separated by a distance of one foot, that have come from millions of miles away, transmitted to the earth from the Mariner II space craft which has just passed Venus. Its signals carry summaries of information it has picked up about the planets (information obtained from electromagnetic waves that travelled from the planet to the space craft).

There are very tiny wiggles of the electric and magnetic fields that are waves which originated billions of light years away—from galaxies in the remotest corners of the universe. That this is true has been found by ā€œfilling the room with wiresā€ā€”by building antennas as large as this room. Such radiowaves have been detected from places in space beyond the range of the greatest optical telescopes. Even they, the optical telescopes, are simply gatherers of electromagnetic waves. What we call the stars are only inferences, inferences drawn from the only physical reality we have yet gotten from them—from a careful study of the unendingly complex undulations of the electric and magnetic fields reaching us on earth.

There is, of course, more: the fields produced by lightning miles away, the fields of the charged cosmic ray particles as they zip through the room, and more, and more. What a complicated thing is the electric field in the space around you! Yet it always satisfies the three-dimensional wave equation.”

So… Well… That’s it for today, folks. šŸ™‚ We have some more gymnastics to do, still… But we’re really there. Or here, I should say: on top of the peak. What a view we have here! Isn’t it beautiful? It took us quite some effort to get on top of this thing, and we’re still trying to catch our breath as we struggle with what we’ve learned so far, but it’s really worthwhile, isn’t it? šŸ™‚

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 20, 2020 as a result of a DMCA takedown notice from Michael A. Gottlieb, Rudolf Pfeiffer, and The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 20, 2020 as a result of a DMCA takedown notice from Michael A. Gottlieb, Rudolf Pfeiffer, and The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/

Maxwell’s equations and the speed of light

Pre-script (dated 26 June 2020): Our ideas have evolved into a full-blown realistic (or classical) interpretation of all things quantum-mechanical. In addition, I note the dark force has amused himself by removing some material, which messed up the lay-out of this post as well. So no use to read this. Read my recent papers instead. šŸ™‚

Original post:

We know how electromagnetic waves travel through space: they do so because of the mechanism described in Maxwell’s equation: a changing magnetic field causes a changing electric field, and a changing magnetic field causes a (changing) electric field, as illustrated below.

Maxwell interaction

So we needĀ someĀ First Cause to get it all startedĀ šŸ™‚ i.e. some current, i.e. some moving charge, but then the electromagnetic wave travels, all by itself, through empty space, completely detached from the cause. You know that by now – indeed, you’ve heard this a thousand times before – but, if you’re reading this, you want to know how it worksĀ exactly. šŸ™‚

In my post on the Lorentz gauge, I included a few links to Feynman’sĀ LecturesĀ that explain the nitty-gritty of this mechanism from various angles. However, they’re pretty horrendous to read, and so I just want to summarize them a bit—if only for myself, so as to remind myself what’s important and not. In this post, I’ll focus on the speed of light: why do electromagnetic waves – light – travel at the speed of light?

You’ll immediately say: that’s a nonsensical question. It’s light, so it travels at the speed of light. Sure, smart-arse!Ā Let me be more precise: how can we relate the speed of light to Maxwell’s equations? That is the question here. Let’s go for it.

Feynman deals with the matter of the speed of an electromagnetic wave, and the speed of light, in a rather complicated exposé on the fieldsĀ from someĀ infinite sheet of charge that is suddenly set into motion, parallel to itself, as shown below. The situation looks – and actually is – very simple, but the math is rather messy because of the rather exotic assumptions: infinite sheets and infinite acceleration are not easy to deal with. šŸ™‚ But so the whole point of theĀ exposé is just to proveĀ that the speed of propagation (v) of the electric and magnetic fields is equal to the speed of light (c), and it does a marvelous job at that. So let’s focus on that hereĀ only. So what I am saying is that I am going to leave out most of the nitty-gritty and just try to get to that v = cĀ result as fast as I possibly can. So, fasten your seat belt, please.

sheet of charge

Most of the nitty-gritty in Feynman’sĀ exposé is about how to determine the direction and magnitude of the electric and magnetic fields, i.e. E and B. Now, when the nitty-gritty business is finished, the grand conclusion is that both E and B travel out in both the positive as well as the negative x-direction at some speed vĀ and sort of ‘fill’ the entire space as they do. Now, the regionĀ they are filling extends infinitely far in both the y- and z-direction but, because they travel along the x-axis, there are no fields (yet) in the region beyond x = ± vĀ·t (t = 0 is the moment when the sheet started moving, and it moves in the positive y-direction). As you can see, the sheet of charge fills the yz-plane, and the assumption is that its speed goes from zero to u instantaneously, or very very quickly at least. So the E and B fields move out like a tidal wave, as illustrated below, and thereby ‘fill’ the space indeed, as they move out.

tidal wave

The magnitude of E and B is constant, but it’s not the same constant, and part of the exercise here is to determine the relationship between the two constants. As for their direction, you can see it in the first illustration: B points in the negative z-direction for x > 0 and in the positive z-direction for x < 0, while E‘s directionĀ is opposite to u‘s directionĀ everywhere, so EĀ points in the negative y-direction. As said, you should just take my word for it, because the nitty-gritty on this – which we doĀ notĀ want to deal with here – is all in Feynman and so I don’t want to copy that.

The crux of the argument revolves around what happens at the wavefront itself, as it travels out. Feynman relates flux and circulation there. It’s the typical thing to do: it’sĀ at the wavefront itselfĀ that the fieldsĀ change: before they were zero, and now they are equal to that constant. The fields do notĀ change anywhere else, so there’s no changing flux or circulation business to be analyzed anywhere else.Ā So we define two loops at the wavefront itself: Ī“1Ā and Ī“2. They are normal to each other (cf. the top and side view of the situation below), because the E and B fields are normal to each other. And so then we use Maxwell’s equations to check out what happens with the flux and circulation there and conclude what needs to be concluded. šŸ™‚

top view side view

We start with rectangle Ī“2. So one side isĀ in the region where there are fields, and one side is in the region where the fields haven’t reached yet. There is some magnetic flux through this loop, and it is changing, so there is an emf around it, i.e. some circulation of E. The flux changes because the area in which B exists increases at speed v.Ā Now, the time rate of change of the flux is, obviously, the width of the rectangle L times the rate of change ofĀ the area, so that’s (BĀ·LĀ·vĀ·Ī”t)/Ī”t = BĀ·LĀ·v, with Ī”t some differential time interval co-definingĀ how slow or how fast the field changes. Now, according to Faraday’s Law (see my previous post),Ā this will be equal to minus the line integral ofĀ EĀ around Ī“2, which is EĀ·L. So EĀ·L = BĀ·LĀ·v and, hence, we find:Ā E = vĀ·B.

Interesting! To satisfy Faraday’s equation (which is just one of Maxwell’s equations in integral rather than in differential form), E must equal B times v, with v the speed of propagation of our ‘tidal’ wave. Now let’s look at Ī“1. There we should apply:

IntegralNow the line integral is just BĀ·L, and the right-hand side is EĀ·LĀ·v, so, not forgetting that c2Ā in front—i.e. the squareĀ of the speed of light, as you know!—we get:Ā c2B =Ā EĀ·v, or E = (c2/v)Ā·B.Ā 

Now, the E = vĀ·B and E = (c2/v)Ā·B equations mustĀ bothĀ apply (we’re talking one wave and one and the same phenomenon) and, obviously, that’sĀ onlyĀ possible if v =Ā c2/v, i.e. if v = c. So the wavefront mustĀ travel at the speed of light! Waw !Ā That’s fast. šŸ™‚ Yes. […] Jokes aside, that’s the result we wanted here: we justĀ provedĀ that the speed of travel of an electromagnetic wave must beĀ equal to the speed of light.

As an added bonus, we also showed the mechanism of travel. It’s obvious from the equations we used to prove the result: it works through the derivatives of the fields with respect to time, i.e.Ā āˆ‚E/āˆ‚t andĀ āˆ‚B/āˆ‚t.

Done!Ā Great! Enjoy the view!

Well… Yes and no. If you’re smart, you’ll say: we got this result because of the c2Ā factor in that equation, so Maxwell had already put it in, so to speak. Waw! You really areĀ aĀ smart-arse, aren’t you? šŸ™‚

The thing is… Well… The answer is: no. Maxwell didĀ notĀ put it in. Well… Yes and no. Let me explain. Maxwell’s firstĀ equation was the electric flux lawĀ āˆ‡Ā·E = σ/ε0: the flux of E through a closed surface is proportional to the charge inside. So that’s basically an other way of writingĀ Coulomb’s Law,Ā and ε0Ā was just some constant in it, the electric constant. So it’s a constant of proportionality that depends on the unit in which we measure electric charge. The only reason that it’s there is to make the units come out alright, so if we’d measure charge not in coulombĀ (C)Ā in a unit equal to 1 C/ε0, it would disappear. If we’d do that, our new unit would be equivalent to the charge of some 700,000 protons. You can figure that magical number yourself by checking the values of the proton charge and ε0. šŸ™‚

OK. And then Faraday came up with the exactĀ laws for magnetism, and they involved current and some other constant of proportionality, and Maxwell formalized that by writing āˆ‡Ć—BĀ = μ0j, with μ0Ā the magneticĀ constant. It’s not a flux law but a circulation law: currents cause circulationĀ of B. We get the flux rule from it by integrating it. But currents are movingĀ charges, and so Maxwell knew magnetism was related to the same thing: electric charge. So Maxwell knew the two constants had to be related. In fact, when putting the full set of equations together – there are four, as you know – Maxwell figured out that μ0Ā times ε0Ā would haveĀ toĀ be equal to the reciprocal of c2, with cĀ the speed of propagation of the wave. So Maxwell knew that, whatever the unit of charge, we’d get two constants of proportionality, and electricĀ and aĀ magneticĀ constant, and that μ0·ε0Ā would be equal to 1/c2. However,Ā while he knew that,Ā at the time, light and electromagnetism were considered to be separate phenomena, and so Maxwell did not say that cĀ was the speed of light: the only thing his equations told him was thatĀ cĀ is the speed of propagation of thatĀ ‘electromagnetic’ wave that came out of his equations.

The rest is history.Ā In 1856, the great Wilhelm Eduard Weber – you’ve seen his name before, didn’t you? – did a whole bunch of experiments which measured the electric constant rather precisely, and Maxwell jumped on it and calculated all the rest, i.e. μ0, and so then he took the reciprocal of the square root of μ0·ε0Ā and – Bang! – he hadĀ c, the speed of propagation of the electromagnetic wave he was thinking of. Now, cĀ was some value of the order of 3Ɨ108Ā m/s, and so thatĀ happenedĀ to be the same as the speed of light, which suggested that Maxwell’sĀ c and the speed of light wereĀ actually one and the same thing!

Now, I am a smart-arse too šŸ™‚ and, hence, when I first heard this story, I actually wondered how Maxwell could possibly know the speed of light at the time: Maxwell died many years beforeĀ the Michelson-Morley experiment unequivocally established the value of the speed of light. [In case, you wonder: the Michelson-Morley experiment was done in 1887. So I check it. The fact is that the Michelson-Morley experiment concluded that the speed of light was anĀ absoluteĀ value and that, in the process of doing so, they got a rather precise value for it, but the value of cĀ itself has already been established, more or less, that is, by aĀ Danish astronomer, Ole Rƶmer, in 1676 ! He did so by carefully observing the timing of the repeating eclipses of Io, one of Jupiter’s moons. Newton mentioned his results in his Principia, which he wrote in 1687, dulyĀ noting that it takes about seven to eightĀ minutes for light to travel from the Sun to the Earth. Done! The whole story is fascinating, really, so you should check it outĀ yourself. šŸ™‚

In any case, to make a long story short, Maxwell was puzzled by this mysterious coincidence, but he was bold enough to immediately point to the right conclusion, tentativelyĀ at least, and so he told the Cambridge Philosophical Society, in the very same year, i.e. 1856, that “we can scarcely avoid the inference that light consists in the transverse undulations of the same medium which is the cause of electric and magnetic phenomena.”

So… Well… Maxwell still suggests light needs some medium here, so the ‘medium’ is a reference to the infamous aetherĀ theory, but that’s not the point: what he says here is what we all take for granted now: light is an electromagnetic wave.Ā So now we knowĀ there’s absolute no reason whatsoever to avoid the ‘inference’, but… Well… 160 years ago, it was quite a big deal to suggest something like that. šŸ™‚

So that’s the full story. I hoped you like it. Don’t underestimate what you just did: understanding an argument like this is like “climbing a great peak”, as Feynman puts it. So it is “a great moment” indeed. šŸ™‚Ā The only thing left is, perhaps, to explain the ‘other’ flux rules I used above. Indeed, you know Faraday’s Law:

emf

But that other one? Well… As I explained in my previous post, Faraday’s Law is the integralĀ form of Maxwell’s second equation:Ā āˆ’āˆ‚B/āˆ‚t =Ā āˆ‡Ć—E. The ‘other’ flux rule above – so that’s the one with the c2Ā in front and without a minus sign, is the integral form of Maxwell’s fourth equation: c2āˆ‡Ć—BĀ = j/ε0Ā +Ā āˆ‚E/āˆ‚t, taking into account that we’re talking a wave traveling in free space, so there are no charges and currents (it’s just a wave in empty space—whatever that means) and, hence, the Maxwell equation reduces to c2āˆ‡Ć—BĀ = āˆ‚E/āˆ‚t. Now, I could take you through the same gymnastics as I did in my previous post but, if I were you, I’d just apply the general principle that ”the same equations must yield the same solutions” and so I’d just switch E for B and vice versa in Faraday’s equation. šŸ™‚

So we’re done… Well… Perhaps one more thing. We’ve got these flux rules above telling us that theĀ electromagnetic wave will travel all by itself, through empty space, completely detached from its First Cause. But… […] Well… Again you may think there’s some trick here. In other words, you may think the wavefront has to remain connected to the First Cause somehow, just like the whip below is connected to some person whipping it. šŸ™‚

Bullwhip_effect

There’s no such connection. The whip is not needed. šŸ™‚ If we’d switch off the First Cause after some time T, so our moving sheet stops moving, then we’d have theĀ pulseĀ below traveling through empty space. As Feynman puts it: “The fields have taken off: they are freely propagating through space, no longer connected in any way with the source. The caterpillar has turned into a butterfly!Ā 

wavefront

Now, the last question is always the same: whatĀ areĀ those fields? What’s theirĀ reality? Here, I should refer you to one of the most delightful sections in Feynman’s Lectures. It’s on theĀ scientific imagination. I’ll just quote the introduction to it, but I warmly recommend you go and check it out for yourself: it hasĀ no formulasĀ whatsoever, and so you should understandĀ all of itĀ without any problem at all. šŸ™‚

“I have asked you to imagine these electric and magnetic fields. What do you do? Do you know how? How do I imagine the electric and magnetic field? What do I actually see? What are the demands of scientific imagination? Is it any different from trying to imagine that the room is full of invisible angels? No, it is not like imagining invisible angels. It requires a much higher degree of imagination to understand the electromagnetic field than to understand invisible angels. Why? Because to make invisible angels understandable, all I have to do is to alter their properties a little bit—I make them slightly visible, and then I can see the shapes of their wings, and bodies, and halos. Once I succeed in imagining a visible angel, the abstraction required—which is to take almost invisible angels and imagine them completely invisible—is relatively easy. So you say, ā€œProfessor, please give me an approximate description of the electromagnetic waves, even though it may be slightly inaccurate, so that I too can see them as well as I can see almost invisible angels. Then I will modify the picture to the necessary abstraction.ā€

I’m sorry I can’t do that for you. I don’t know how. I have no picture of this electromagnetic field that is in any sense accurate. I have known about the electromagnetic field a long time—I was in the same position 25Ā years ago that you are now, and I have had 25Ā years more of experience thinking about these wiggling waves. When I start describing the magnetic field moving through space, I speak of the EĀ andĀ BĀ fields and wave my arms and you may imagine that I can see them. I’ll tell you what I see. I see some kind of vague shadowy, wiggling lines—here and there is an E and aĀ BĀ written on them somehow, and perhaps some of the lines have arrows on them—an arrow here or there which disappears when I look too closely at it. When I talk about the fields swishing through space, I have a terrible confusion between the symbols I use to describe the objects and the objects themselves. I cannot really make a picture that is even nearly like the true waves. So if you have some difficulty in making such a picture, you should not be worried that your difficulty is unusual.

Our science makes terrific demands on the imagination. The degree of imagination that is required is much more extreme than that required for some of the ancient ideas. The modern ideas are much harder to imagine. We use a lot of tools, though. We use mathematical equations and rules, and make a lot of pictures. What I realize now is that when I talk about the electromagnetic field in space, I see some kind of a superposition of all of the diagrams which I’ve ever seen drawn about them. I don’t see little bundles of field lines running about because it worries me that if I ran at a different speed the bundles would disappear, I don’t even always see the electric and magnetic fields because sometimes I think I should have made a picture with the vector potential and the scalar potential, for those were perhaps the more physically significant things that were wiggling.

Perhaps the only hope, you say, is to take a mathematical view. Now what is a mathematical view? From a mathematical view, there is an electric field vector and a magnetic field vector at every point in space; that is, there are six numbers associated with every point. Can you imagine six numbers associated with each point in space? That’s too hard. Can you imagine even one number associated with every point? I cannot! I can imagine such a thing as the temperature at every point in space. That seems to be understandable. There is a hotness and coldness that varies from place to place. But I honestly do not understand the idea of a number at every point.

So perhaps we should put the question: Can we represent the electric field by something more like a temperature, say like the displacement of a piece of jello? Suppose that we were to begin by imagining that the world was filled with thin jello and that the fields represented some distortion—say a stretching or twisting—of the jello. Then we could visualize the field. After we ā€œseeā€ what it is like we could abstract the jello away. For many years that’s what people tried to do. Maxwell, AmpĆØre, Faraday, and others tried to understand electromagnetism this way. (Sometimes they called the abstract jello ā€œether.ā€) But it turned out that the attempt to imagine the electromagnetic field in that way was really standing in the way of progress. We are unfortunately limited to abstractions, to using instruments to detect the field, to using mathematical symbols to describe the field, etc. But nevertheless, in some sense the fields are real, because after we are all finished fiddling around with mathematical equations—with or without making pictures and drawings or trying to visualize the thing—we can still make the instruments detect the signals from Mariner II and find out about galaxies a billion miles away, and so on.

The whole question of imagination in science is often misunderstood by people in other disciplines. They try to test our imagination in the following way. They say, ā€œHere is a picture of some people in a situation. What do you imagine will happen next?ā€ When we say, ā€œI can’t imagine,ā€ they may think we have a weak imagination. They overlook the fact that whatever we are allowed to imagine in science must be consistent with everything else we know: that the electric fields and the waves we talk about are not just some happy thoughts which we are free to make as we wish, but ideas which must be consistent with all the laws of physics we know. We can’t allow ourselves to seriously imagine things which are obviously in contradiction to the known laws of nature. And so our kind of imagination is quite a difficult game. One has to have the imagination to think of something that has never been seen before, never been heard of before. At the same time the thoughts are restricted in a strait jacket, so to speak, limited by the conditions that come from our knowledge of the way nature really is. The problem of creating something which is new, but which is consistent with everything which has been seen before, is one of extreme difficulty.”

Isn’t that great? I mean: Feynman, one of the greatest physicists of all time, didn’t write what he wrote above when he was a undergrad student or so. No. He did so in 1964, when he was 45 years old, at the height of his scientific career! And it gets better, because Feynman then starts talking aboutĀ beauty. What is beauty in science? Well… Just click and check what Feynman thinks about it. šŸ™‚

Oh… Last thing. So what is the magnitude of the E and B field? Well… You can work it out yourself, but I’ll give you the answer. The geometry of the situation makes it clear that the electric field has a y-component only, and the magnetic field a z-component only. Their magnitudes are given in terms of J, i.e. the surface current density going in the positive y-direction:

equation

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/

An introduction to electric circuits

In my previous post,I introduced electric motors, generators and transformers. They all work because of Faraday’s flux rule: a changing magnetic flux will produce some circulation of the electric field. The formula for the flux rule is given below:

emf

It is a wonderful thing, really, but not easy to grasp intuitively. It’s one of these equations where I should quote Feynman’s introduction to electromagnetism: “The laws of Newton were very simple to write down, but they had a lot of complicated consequences and it took us a long time to learn about them all. The laws of electromagnetism are not nearly as simple to write down, which means that the consequences are going to be more elaborate and it will take us quite a lot of time to figure them all out.”

Now, among Maxwell’s Laws, this is surely the most complicated one! However, that shouldn’t deter us. šŸ™‚ RecallingĀ Stokes’ Theorem helps to appreciate what the integral on the left-hand side represents:

Stokes theorem

We’ve got a line integral around some closed loopĀ Ī“ on the left and, on the right, we’ve got a surface integral over some surface S whose boundary is Ī“. The illustration below depicts the geometry of the situation. You know what it all means. If not, I am afraid I have to send you back to square one, i.e. my posts on vector analysis. Yep. Sorry. Can’t keep copying stuff and make my posts longer and longer. šŸ™‚

Diagram stokesTo understand the flux rule, you should imagine that the loop Ī“ is some loop of electric wire, and then you just replace C by E, the electric field vector. The circulation of E, which is caused by theĀ changeĀ in magnetic flux, is referred to as the electromotive forceĀ (emf), and it’s theĀ tangential force (EĀ·ds) per unit charge in the wire integrated over its entire length around the loop, which is denoted byĀ Ī“ here, and which encloses a surface S.

Now, you can go from the line integral to the surface integral by noting Maxwell’s Law:Ā āˆ’āˆ‚B/āˆ‚t =Ā āˆ‡Ć—E. In fact, it’s the same flux rule really, but in differential form. As for (āˆ‡Ć—E)n, i.e. the component of āˆ‡Ć—EĀ that is normal to the surface, you know that any vector multiplied with the normal unit vector will yield its normal component. In any case, if you’re reading this, you should already be acquainted with all of this. Let’s explore the concept of the electromotive force, and then apply it our first electric circuit. šŸ™‚

Indeed, it’s now time for a small series on circuits, and so we’ll start right here and right now, but… Well… First things first. šŸ™‚

The electromotive force: concept and units

The term ‘force’ in ‘electromotive force’ is actually somewhat misleading. There is a force involved, of course, but the emf is not a force. The emf is expressed in volts.Ā That’s consistent with its definition as the circulation of E: a force times a distance amounts to work, or energyĀ (oneĀ jouleĀ is one newtonĀ·meter), and because E is the force on aĀ unit charge, the circulation of E is expressed in jouleĀ per coulomb, so that’s a voltage: 1Ā volt = 1 joule/coulomb.Ā Hence, on the left-hand side of Faraday’s equation, we don’t have any dimension of time: it’s energy per unit charge, so it’s xĀ joule per coulombĀ . Full stop.

On the right-hand side, however, we have the time rate of change of the magnetic flux. through the surface S. The magnetic flux is a surface integral, and so it’s a quantity expressed in [B]Ā·m2, with [B] the measurement unit for the magnetic field strength. The time rate of change of the flux is then, of course, expressed in [B]Ā·m2Ā per second, i.e. [B]Ā·m2/s.Ā Now what is the unit for the magnetic field strength B, which we denoted by [B]?

Well… [B] is a bit of a special unit: it is notĀ measured as some force per unit charge, i.e. in newton per coulomb, like the electric field strengthĀ E. No. [B] is measured in (N/C)/(m/s). Why? Because the magnetic force is not F = qE butĀ FĀ = qvƗB. Hence, so as to make the units come out alright, we need to express B in (NĀ·s)/(CĀ·m), which is a unit known as the teslaĀ (1 T = NĀ·s/CĀ·m), so as to honor the Serbian-American genius Nikola Tesla. [I know it’s a bit of short and dumb answer, but the complete answer is quite complicated: it’s got to do with theĀ relativityĀ of the magnetic force, which I explained in another post: both the v in FĀ = qvƗB equationĀ as well as theĀ m/s unit in [B] should make you think: whoseĀ velocity? In which reference frame? But that’s something I can’t summarize in two lines, so just click the link if you want to know more. I need to get back to the lesson.]

Now that we’re talking units, I should note that the unit of flux also got a special name, theĀ weber, so as to honor one of Germany’s most famous physicists, Wilhelm Eduard Weber: as you might expect, 1 Wb = 1 TĀ·m2. But don’t worry about these strange names. Besides the units you know, like the jouleĀ and theĀ newton,Ā I’ll only use the volt,Ā which got its name to honor some other physicist, Alessandro Volta, the inventor of the electrical battery. Or… Well… I might mention theĀ wattĀ as well at some point… šŸ™‚

So how does it work? On one side, we have something expressedĀ per second – so that’s per unit time – and on the other we have something that’s expressedĀ per coulomb – so that’s per unit charge. The link between the two is theĀ power, so that’s theĀ time rate of doing work. It’s expressed inĀ joule per second. So… Well… Yes. Here we go: in honor of yet another genius, James Watt, the unit of power got its own special name too: theĀ watt. šŸ™‚ In the argument below, I’ll show that the power that is beingĀ generatedĀ by a generator, and that is beingĀ consumedĀ in the circuit (through resistive heating, for example, or whatever else taking energy outĀ of the circuit) is equal to theĀ emf times the current. For the moment, however, I’ll just assume you believe me. šŸ™‚

We need to look at the whole circuit now, indeed, in which our little generator (i.e. our loop or coil of wire) is just one of the circuit elements. The units come out alright: the power =Ā emfĀ·currentĀ product is expressed in voltĀ·coulomb/second = (joule/coulomb)Ā·(coulomb/second) = joule/second. So, yes, it looks OK. But what’s going on really? How does it work, literally?

A short digression: on Ohm’s Law and electric power

Well… Let me first recall the basic concepts involved which, believe it or not, are probably easiest to explain by briefly recalling Ohm’s Law, which you’ll surely remember from your high-school physics classes. It’s quite simple really: we have someĀ resistanceĀ in a little circuit, so that’s something thatĀ resistsĀ the passage of electric current, and then we also have a voltage source. Now, Ohm’s Law tells us that the ratio of (i) theĀ voltage VĀ across the resistance (so that’s between the two points marked as + andĀ āˆ’) and (ii) the currentĀ I will be some constant. It’s the same as saying that V and I are inversely proportional to each other.Ā  The constant of proportionality is referred to as theĀ resistanceĀ itself and, while it’s often looked at as a property of the circuit itself, we may embody it in a circuit element itself: a resistor, as shown below.

120px-OhmsLaw

So we write R = V/I, and the brief presentation above should remind you of theĀ capacityĀ of a capacitor, which was just another constant of proportionality. Indeed, instead of feeding a resistor (so all energyĀ gets dissipated away),Ā we could charge a capacitor with a voltage source, so that’s a energyĀ storageĀ device, and thenĀ we find that the ratio between (i) the charge on the capacitor and (ii) the voltage across the capacitor was a constant too, which we defined as the capacity of the capacitor, and so we wrote C = Q/V. So, yes, another constant of proportionality (there are many in electricity!).

In any case, the point is: to increase the current in the circuit above, you need to increase the voltage, but increasing both amounts to increasing the powerĀ that’s being consumed in the circuit, because theĀ power is voltage times current indeed, so P = VĀ·I (or vĀ·i, if I use the small letters that are used in the two animations below). For example, if we’d want to double the current, we’d need to double the voltage, and so we’re quadrupling the power: (2Ā·V)Ā·(2Ā·I) = 22Ā·VĀ·I. So we have a square-cube law for the power, which we get by substituting V for RĀ·I or by substituting I for V/R, so we can write the power P as P = V2/R = I2Ā·R. This square-cube law says exactly the same: if you want to doubleĀ the voltageĀ orĀ the current, you’ll actually have to double both and, hence, you’ll quadruple the power.Ā Now let’s look at the animations below (for which credit must go to Wikipedia).

Electric_power_source_animation_1 Electric_load_animation_2

TheyĀ show how energy is being used in an electric circuit in Ā terms of power. [Note that the little moving plusesĀ are in line with the conventionĀ that a current is defined as the movement of positive charges, so we write I = dQ/dt instead ofĀ I =Ā āˆ’dQ/dt. That also explains the direction of the field line E, which has been added to show that the power source effectively moves charges againstĀ the field and, hence, against the electric force.] What we have here is that, on one side of the circuit, someĀ generator or voltage sourceĀ will create anĀ emfĀ pushing the charges,Ā and then someĀ loadĀ willĀ consumeĀ their energy, so they lose their push. SoĀ power, i.e. energy per unit time, is supplied, and is then consumed.

Back to the emf…

Now, I mentioned that theĀ emf is a ratio of two terms: the numerator is expressed inĀ joule, and the denominator is expressed inĀ coulomb. So you might think we’ve got some trade-off here—something like: if weĀ doubleĀ the energy of halfĀ ofĀ the individual charges, then we still get the same emf. Or vice versa: we could, perhaps, double the number of charges and load them with only half the energy. One thing is for sure: we can’t both.

Hmm… Well… Let’s have a look at this line of reasoning by writing it down more formally.

  1. The time rate of change of the magnetic flux generates some emf, which we can and should think of as a property of the loop or the coil of wire in which it is being generated. Indeed, the magnetic flux through it depends on its orientation, its size, and its shape. So it’s really very much like the capacity of a capacitor or the resistance of a conductor. SoĀ we write: emf = Ī”(flux)/Ī”t. [In fact, the inducedĀ emfĀ tries to oppose the change in flux, so I should add the minus sign, but you get the idea.]
  2. For a uniform magnetic field, the flux is equal to the field strength B times the surface area S. [To be precise, we need to take the normal component of B, so the flux is BĀ·S = BĀ·SĀ·cosĪø.] Ā So the flux can change because of a change in B or because of a change in S, or because of both.
  3. The emf = Ī”(flux)/Ī”t formula makes it clear that a veryĀ slowĀ change in flux (i.e. the same Ī”(flux) over a much larger Ī”t) will generate little emf. In contrast, a very fastĀ change (i.e. the the same Ī”(flux) over a much smaller Ī”t) will produce a lot of emf. So, in that sense, emf is notĀ like the capacity or resistance, because it’s variable: it depends on Ī”(flux), as well as on Ī”t. However, you should still think of it as aĀ property of the loop or the ‘generator’ we’re talking about here.
  4. Now, the power that is being produced or consumed in the circuit in which our ‘generator’ is just one of the elements, is equal to the emfĀ times the current. The power is the time rate of change of the energy, and the energy is the workĀ that’s being done in the circuit (which I’ll denote by Ī”U), so we write: emfĀ·current =Ā Ī”U/Ī”t.
  5. Now, the current is equal to the time rate of change of the charge, so I = ΔQ/Δt. Hence, the emf is equal to emf = (ΔU/Δt)/I = (ΔU/Δt)/(ΔQ/Δt) = ΔU/ΔQ. From this, it follows that: emf = Δ(flux)/Δt = ΔU/ΔQ, which we can re-write as:

Δ(flux) = ΔU·Δt/ΔQ

What this says is the following. For a given amount of change in the magnetic flux (so we treatĀ Ī”(flux) as constant in the equation above), we could do more work on the same charge (Ī”Q) – we could double Ī”U by moving the same charge over a potential difference that’s twice as large, for example – but then Ī”t must be cut in half. So the same change in magnetic flux can doĀ twice as much work if the change happens in half of the time.

Now, does that mean the current is being doubled? We’re talking the same Ī”Q and half the Ī”t, so… Well? No. The Ī”t here measures the time of the flux change, so it’s not the dt in I = dQ/dt. For the current to change, we’d need to move the same charge faster, i.e. over a larger distance over the same time. We didn’t say we’d do that above: we only said we’d move the charge across a larger potential difference: we didn’t say we’d change the distanceĀ over which they are moved.

OK. That makes sense. But we’re not quite finished. Let’s first try something else, to then come back to where we are right now via some other way. šŸ™‚ Can we change Ī”Q? Here we need to look at the physics behind. What’s happening really is that the change in magnetic flux causes an induced current which consists of the free electrons in the Γ loop. So we have electrons moving in and out of our loop, and through the whole circuit really, but so there’s only so many free electrons per unit lengthĀ in the wire. However, if we would effectively double the voltage, then theirĀ speedĀ will effectively increase proportionally, so we’ll have more of them passing throughĀ per second. Now that effect surely impacts the current. It’s what we wrote above: all other things being the same, including the resistance, then we’ll also double the currentĀ as we double the voltage.

So where is thatĀ effect in the flux rule? The answer is: it isn’t there. The circulation of E around the loop is what it is: it’s some energy per unit charge. Not per unit time. So our flux rule gives us a voltage, which tells us that we’re going to have some pushĀ on the charges in the wire,Ā but it doesn’t tell us anything about the current. To know the current, we must know theĀ velocityĀ of the moving charges, which we can calculateĀ from the push if we also get some other information (such as the resistance involved, for instance), but so it’s not there in the formula of the flux rule. You’ll protest: there is a Ī”t on the right-hand side! Yes, that’s true. But it’s not theĀ Ī”t in the v =Ā Ī”s/Ī”t equation for our charges. Full stop.

Hmm… I may have lost you by now. If not, please continue reading. Let me drive the point home by asking another question. Think about the following: we can re-write thatĀ Ī”(flux) = Ī”UĀ·Ī”t/Ī”Q equation above asĀ Ī”(flux) = (Ī”U/Ī”Q)Ā·Ī”t equation. Now, does that imply that, withĀ the same change in flux, i.e. the same Ī”(flux), and, importantly, for the same Ī”t, we could double both Ī”U as well as Ī”Q? I mean: (2Ā·Ī”U)/(2Ā·Ī”Q) = Ī”U/Ī”Q and so the equation holds, mathematicallyĀ that is. […] Think about it.

You should shake your head now, and rightly so, because, while theĀ Ī”(flux) = (Ī”U/Ī”Q)Ā·Ī”t equation suggests that would be possible, it’s totally counter-intuitive. We’re changing nothing in the real world (what happens there is the same change of flux in the same amount of time), but so we’d get twice the energy and twice the charge ?! Of course, we could also put a 3 there, or 20,000, or minus a million. So who decides on what we get? You get the point: it is, indeed, notĀ possible. Again, what we can change is the speed of the free electrons, but not their number, and to change their speed, you’ll need to do more work, and so the realityĀ is thatĀ we’re always looking at the same Ī”Q, so if we want a larger Ī”U, then we’ll need a larger change in flux, or we a shorter Ī”t during which that change in flux is happening.

So what can we do? We can change the physics of the situation. We can do so in many ways, like we could change the length of the loop, or its shape. One particularly interesting thing to do would be toĀ increase the number of loops, so instead of one loop, we could have someĀ coil with, say, N turns, so that’s N of these Ī“ loops. So what happens then? In fact, contrary to what you might expect, theĀ Ī”Q still doesn’t change as it moves into the coil and then from loop to loop to get out and then through the circuit: it’s still the same Ī”Q. But the work that can be done by this current becomes much larger. In fact, two loops give us twiceĀ the emf of one loop, and N loops give us N times the emf of one loop. So then we can make the free electrons move faster, so they cover more distance in the same time (and you know work is force times distance), or we can move them across a larger potential difference over the same distance (and so then we move them against a larger force, so it also implies we’re doing more work). The first case is a larger current, while the second is a larger voltage. So what is it going to be?

Think about the physics of the situation once more: to make the charges move faster, you’ll need a larger force, so you’ll have a larger potential difference, i.e. a larger voltage. As for what happens to the current, I’ll explain that below. Before I do, let me talk some more basics.

In the exposĆ© below, we’ll talk about power again, and also about load. What is load? Think about what it is in real life: when buying a battery for a big car, we’ll want a big battery, so we don’t look at the voltage only (they’re all 12-volt anyway). We’ll look at howĀ manyĀ ampĆØres it can deliver, and for how long. The starter motorĀ in the car, for example, can suck up like 200 A, but for a very short time only, of course, as the car engine itself should kick in. So that’s why the capacity of batteries is expressed in ampĆØre-hours.

Now, how do we get such large currents, such large loads? Well… Use Ohm’s Law: to get 200 A at 12 V, the resistance of the starter motor will have to as low as 0.06 ohm. So large currents are associated with very lowĀ resistance. Think practical: a 240-volt 60 wattĀ light-bulb will suck in 0.25 A, and hence, its internal resistance, is about 960 Ī©. Also think of what goes on in your house: we’ve got a lot of resistors in parallel consuming power there. The formula for the total resistance is 1/RtotalĀ =Ā 1/R1Ā + 1/R2Ā + 1/R3Ā + … So more appliances is lessĀ resistance, so that’s what draws in the larger current.

The point is: when looking at circuits, emfĀ is one thing, but energy and power, i.e. the work done per second, are all that matters really. And so then we’re talking currents, but our flux rule does notĀ say how much current our generator will produce: that depends on the load. OK. We really need to get back to the lesson now.

A circuit with an AC generator

The situation is depicted below. We’ve got a coil of wire of, let’s say, N turns of wire, and we’ll use it to generate an alternating currentĀ (AC) in a circuit.

AC generatorCircuit

The coil is really like the loop of wire in that primitive electric motor I introduced in my previous post, but so now we use the motor as a generator. To simplify the analysis, we assume we’ll rotate our coil of wire in aĀ uniformĀ magnetic field, as shown by the field lines B.

motor

Now, our coil is not a loop, of course: the two ends of the coil are brought to external connections through some kind of sliding contacts, but that doesn’t change the flux rule: a changingĀ magnetic fluxĀ will produce someĀ emfĀ and, therefore, someĀ currentĀ in the coil.

OK. That’s clear enough. Let’s see what’s happening really. When we rotate our coil of wire, we change the magnetic flux through it. If S is the area of the coil, andĀ Īø is the angle between the magnetic field and the normal to the plane of the coil, then the flux through the coil will be equal to BĀ·SĀ·cosĪø. Now, if we rotate the coil at a uniform angular velocity ω, then Īø varies with time as Īø = ω·t. Now, each turn of the coil will have an emf equal to the rate of change of the flux, i.e. d(BĀ·SĀ·cosĪø)/dt. We’ve got N turns of wire, and so the total emf, which we’ll denote by Ɛ (yep, a new symbol), will be equal to:

Formula emfNow, that’s just a nice sinusoidal function indeed, which will look like the graph below.

graph (1)

When no current is being drawn from the wire, this Ɛ will effectively be the potential difference between the two wires. What happens really is that theĀ emfĀ produces a current in the coil which pushes some charges out to the wire, and so then they’re stuck there for a while, and so there’s aĀ potential differenceĀ between them, which we’ll denote by V, and that potential difference will be equal to ʐ. It hasĀ to be equal to ʐ because, if it were any different, we’d have an equalizing counter-current, of course. [It’s a fine point, so you should think about it.] So we can write:

formula VSo what happens when weĀ doĀ connect the wires to the circuit, so we’ve got that closed circuitĀ depicted above (and below)?

Circuit

Then we’ll have a current I going through the circuit, and Ohm’s Law then tells us that the ratio between (i) theĀ voltage across the resistance in this circuit (we assume the connections between the generator and the resistor itself are perfect conductors) and (ii) the current will be some constant, so we have R = V/I and, therefore:

Formula AC generator

[To be fully complete, I should note that, when otherĀ circuit elements than resistors are involved, like capacitors andĀ inductors, we’ll have aĀ phase difference between the voltage and current functions, and so we should look at the impedanceĀ of the circuit, rather than its resistance. For more detail, see the addendum below this post.]

OK. Let’s now look at the power and energy involved.

Energy and power in the AC circuit

You’ll probably have many questions about the analysis above. You should. I do. The most remarkable thing, perhaps, is that this analysis suggests that the voltage doesn’t drop as we connect the generator to the circuit. It should. Why not? Why do the charges at both ends of the wire simplyĀ dischargeĀ through the circuit? In real life, there surely is such tendency:Ā sudden large changes in loading will effectively produce temporary changes in the voltage. But then it’s like FeynmanĀ writes: “The emf will continue to provide charge to the wires as current is drawn from them, attempting to keep the wires always at the same potential difference.”

So how much current is drawn from them? As I explained above, that dependsĀ notĀ on the generator but on the circuit, and more in particular on theĀ load, so that’s the resistor in this case. Again, the resistance is the (constant) ratio of the voltage and the current: R = V/I. So think about increasing or decreasing the resistance. If theĀ voltageĀ remains the same, it implies the current must decrease or increase accordingly, because R = V/I implies that I = V/R. So the current is inversely proportionalĀ to R, as I explained above when discussing car batteries and lamps and loads. šŸ™‚

Now, I still have to prove that theĀ powerĀ provided by our generator is effectively equal to P = Ɛ·I but, if it is, it implies the power that’s being delivered will be inversely proportional to R. Indeed, when ʐ and/or V remain what they are as we insert a larger resistance in the circuit, then P = Ɛ·I = ʐ2/R, and so the power that’s being delivered would be inversely proportional to R. To be clear, we’d have a relation between P and R like the one below.

Capture

This is somewhat weird. Why? Well… I also have to show you that the power that goes into moving our coil in the magnetic field, i.e. the rate ofĀ mechanicalĀ work required to rotate the coil against the magnetic forces, is equal to theĀ electricĀ power Ɛ·I, i.e. the rate at which electrical energy is being delivered by the emf of the generator. However, I’ll postpone that for a while and, hence, I’ll just ask you, once again, to take me on my word. šŸ™‚ Now, if that’s true, so if the mechanical power equals the electric power, then that implies that a larger resistance will reduce theĀ mechanical power we need to maintain the angular velocity ω. Think of a practical example: if we’d double the resistance (i.e. we halve the load), and if the voltage stays the same, then the current would be halved, and the power would also be halved. And let’s think about the limit situations: as the resistance goes to infinity, the power that’s being delivered goes to zero, as the current goes to zero, while if the resistance goes to zero, both the current as well as the power would go to infinity!

Well… We actuallyĀ know that’s alsoĀ true in real-life: actual generators consume more fuel when the load increases, so when they deliver more power,Ā and much less fuel, so less power, when there’s no load at all. You’ll know that, at least when you’re living in a developing country with a lot of load shedding! šŸ™‚ And the difference is huge: no or just a little load will only consume 10% of what you need when fully loading it. It’s totally in line with what I wrote on the relationship between the resistance and the current that it draws in. So, yes, it does make sense:

AnĀ emf doesĀ produceĀ more current if the resistance in the circuitĀ is lowĀ (so i.e. when the load is high), and the stronger currents do represent greater mechanical forces.

That’s a veryĀ remarkable thing. It means that, if we’d put a largerĀ load on our little AC generator, it should require more mechanical workĀ to keep the coil rotating at the same angular velocity ω. But… What changes? The change in flux is the same, theĀ Ī”t is the same, and so what changes really?Ā What changes is the current going through the coil, and it’s not a change in that Ī”Q factor above, but a change in its velocity v.

Hmm… That all looks quite complicated, doesn’t it? It does, soĀ let’s get back to the analysis of what we have here, so we’ll simply assume that we have someĀ dynamic equilibriumĀ obeying that formula above, and so I and R are what they are, and we relate them to ʐ according to that equation above, i.e.:

Formula AC generator

Now let me prove those formulas on the powerĀ of our generator and in the circuit. We have all these charges in our coil that are receiving some energy. Now, theĀ rateĀ at which they receive energy is FĀ·v.

Huh? Yes.Ā Let me explain: theĀ workĀ that’s being done on a charge along some path is the line integral ∫ FĀ·ds along this path. But the infinitesimal distance ds is equal toĀ vĀ·dt, as ds/dt = vĀ (note that we write s and v as vectors, so the dot product with F gives us the component of F that is tangential to the path). So ∫ FĀ·ds = ∫ (FĀ·v)dt. So the time rate of change of the energy, which is the power, isĀ FĀ·v. Just take the time derivative of the integral. šŸ™‚

Now let’s assume we have n moving charges per unit length of our coil (so that’s in line with what I wrote about Ī”Q above), then the power being delivered to any element ds of the coil is (FĀ·v)Ā·nĀ·ds, which can be written as:Ā (FĀ·ds)Ā·nĀ·v. [Why? Because v and ds have the same direction: the direction of both vectors is tangential to the wire, always.] Now all we need to do to find out how much power is being delivered to the circuit by our AC generatorĀ is integrate this expression over the coil, so we need to find:

power

However, the emf (Ɛ) is defined as the line integral ∫ E·ds line, taken around the entire coil, and E = F/q, and the current I is equal to I = q·n·v. So the power from our little AC generator is indeed equal to:

Power = Ɛ·I

So that’s done. Now I need to make good on my other promise, and that is to show that Ɛ·I product is equal to theĀ mechanicalĀ power that’s required to rotate the coil in the magnetic field. So how do we do that?

We know there’s going to be some torque because of the current in the coil. It’s formula is given byĀ Ļ„ = μ×B. What magnetic field? Well… Let me refer you to my post on the magnetic dipole and its torque: it’sĀ notĀ the magnetic fieldĀ causedĀ by the current, but the external magnetic field, so that’s the B we’ve been talking about here all along. So… Well… I am not trying to fool you here. šŸ™‚ However, the magnetic moment μ was notĀ defined by that external field, but by the current in the coil and its area. Indeed, μ‘s magnitude was the current times the area, so that’s NĀ·IĀ·S in this case. Of course, we need to watch out because μ is a vector itself and so we need the angle between μ andĀ B to calculate that vector cross productĀ Ļ„ = μ×B. However, if you check how we defined the direction of μ, you’ll see it’s normal to the plane of the coil and, hence, the angle between μ andĀ B is the very same ĪøĀ = ω·t that we started our analysis with. So, to make a long story short, theĀ magnitudeĀ of the torqueĀ Ļ„ is equal to:

Ļ„ = (NĀ·IĀ·S)Ā·BĀ·sinĪø

Now, we know the torque is also equal to theĀ work done per unit of distance traveledĀ (around the axis of rotation, that is), soĀ Ļ„ = dW/dĪø. Now dĪø = d(ω·t) = ω·dt. So we can now find the work done per unit of time, so that’s the powerĀ once more:

dW/dt = ω·τ = ω·(NĀ·IĀ·S)Ā·BĀ·sinĪø

But so we found that ʐ = NĀ·SĀ·B·ω·sinĪø, so… Well… We find that:

dW/dt = Ɛ·I

Now, this equation doesn’t sort out our question as to howĀ muchĀ power actually goes in and out of the circuit as we put someĀ loadĀ on it, but it is what we promised to do:Ā I showed that the mechanical work we’re doing on the coil is equal to theĀ electricĀ energy that’s being delivered to the circuit. šŸ™‚

It’s all quite mysterious, isn’t it? It is. And we didn’t include other stuff that’s relevant here, such as the phenomenon of self-inductance: the varying current in the coil will actually produce its own magnetic field and, hence, in practice, we’d get some “back emf” in the circuit. This “back emf” is opposite to the current when it is increasing, and it is in the direction of the current when it is decreasing.Ā In short, the self-inductance effect causes a current to have ‘inertia’: the inductive effects try to keep the flow constant, just as mechanical inertia tries to keep the velocity of an object constant. But… Well… I left that out. I’ll take about next time because…

[…] Well… It’s getting late in the day, and so I must assume this is sort of ‘OK enough’ as an introduction to what we’ll be busying ourselves with over the coming week. You take care, and I’ll talk to you again some day soon. šŸ™‚

Perhaps one little note, on a question that might have popped up when you were reading all of the above: so how do actualĀ generators keep the voltage up? Well… Most AC generators are, indeed, so-calledĀ constant speed devices. You can download some manuals from the Web, and you’ll find things like this: don’t operate at speeds above 4% of the rated speed, or more than 1% below the rated speed. Fortunately, the so-called engine governorĀ will take car of that. šŸ™‚

Addendum: The concept of impedance

In one of my posts on oscillators, I explain the concept of impedance, which is the equivalent ofĀ resistance, but for AC circuits. Just like resistance, impedance also sort ofĀ measures the ā€˜opposition’ that a circuit presents to a current when a voltage is applied, but it’s a complexĀ ratio, as opposed to R = V/I. It’s literally a complex ratio because the impedance has a magnitude and a direction, or aĀ phaseĀ as it’s usually referred to. Hence, one will often write the impedance (denoted by Z) using Euler’s formula:

Z =Ā |Z|eiĪø

The illustration below (credit goes to Wikipedia, once again) explains what’s going on. It’s a pretty generic view of the same AC circuit. The truth is: if we apply an alternating current, then the current and the voltage will both go up and down, but the current signal will usually lag the voltage signal, and the phase factor Īø tells us by how much. Hence, using complex-number notation, we write:

V = Iāˆ—ZĀ = Iāˆ—|Z|eiĪø

General_AC_circuit

Now, while that resembles the V = RĀ·I formula, you should note the bold-face type for V and I, and theĀ āˆ— symbol I am using here for multiplication. First theĀ āˆ— symbol: that’s to make it clear we’re not talking a vector cross product AƗB here, but a product of two complex numbers. The bold-face for V and I implies they’re like vectors, or like complex numbers: so they have a phase too and, hence, we can write them as:

  • VĀ =Ā |V|ei(ωt +Ā ĪøV)
  • IĀ =Ā |I|ei(ωt +Ā ĪøI)

To be fully complete – you may skip all of this if you want, but it’s not that difficult, nor very long – it all works out as follows. We write:

VĀ =Ā Iāˆ—ZĀ = |I|ei(ωt +Ā ĪøI)āˆ—|Z|eiĪøĀ = |I||Z|ei(ωt +Ā ĪøIĀ +Ā Īø)Ā =Ā |V|ei(ωt +Ā ĪøV)

Now, this equation must hold for all t, so we can equate the magnitudes and phases and, hence, we get:Ā |V| = |I||Z| and so we get the formula we need, i.e. theĀ phase differenceĀ between our function for the voltage and our function for the current.

ĪøVĀ = ĪøIĀ + Īø

Of course, you’ll say: voltage and current are something real, isn’t it? So what’s this about complex numbers? You’re right. I’ve used the complex notation only to simplify the calculus, so it’s only theĀ real partĀ of those complex-valued functions that counts.

Oh… And also note that, as mentioned above, we do not have such lag or phase difference when onlyĀ resistors are involved. So we don’t need the concept of impedance in the analysis above. With this addendum, I just wanted to be as complete as I can be. šŸ™‚

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 17, 2020 as a result of a DMCA takedown notice from Michael A. Gottlieb, Rudolf Pfeiffer, and The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 20, 2020 as a result of a DMCA takedown notice from Michael A. Gottlieb, Rudolf Pfeiffer, and The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/

Induced currents

In my two previous posts, I presented all of the ingredients of the meal we’re going to cook now, most notably:

  1. The formula for theĀ torque on a loop of a current in a magnetic field, and its energy: (i)Ā Ļ„ = μ×B, andĀ (ii)Ā Umech = āˆ’Ī¼Ā·B.
  2. The Biot-Savart Law, which gives you the magnetic field that’s produced by wires carrying currents:

B formula 2

Both ingredients are, obviously, relevant to the design of an electromagneticĀ motor, i.e. an ‘engine that can do some work’, as Feynman calls it. šŸ™‚ Its principle is illustrated below.

motor

The two formulas above explain how and why the coil go around, andĀ theĀ coil can be made to keep going by arranging that the connections to the coil are reversed each half-turn by contacts mounted on the shaft. Then the torque is always in the same direction. That’s how a small direct currentĀ (DC) motor is made. MyĀ father made me make a couple of these thirty years ago, with a magnet, a big nail and some copper coil. I used sliding contacts, and they were the most difficult thing in the whole design. But now I found a very nice demo on YouTubeĀ of a guy whose system to ‘reverse’ the connections is wonderfully simple: he doesn’t use any sliding contacts. He just removes half of the insulation on the wire of the coil on one side. It works like a charm, but I think it’s not so sustainable, as it spins so fast that the insulation on the other side will probably come off after a while! šŸ™‚

Now, to make this motor run, you needĀ currentĀ and, hence, 19th century physicists and mechanical engineers also wondered how one could produce currents byĀ changingĀ the magnetic field. Indeed, they could use Alessandro Volta’s ‘voltaic pile‘ to produce currents but it was not very handy: it consisted of alternating zinc and copper discs, withĀ pieces of cloth soaked in salt water in-between!

Now, while theĀ Biot-Savart Law goes back to 1820, it took another decade to find out how that could be done.Ā Initially, people thought magnetic fields should just cause some current, but that didn’t work. Finally, Faraday unequivocally established the fundamental principle thatĀ electric effects are only there when something is changing.Ā So you’ll get a current in a wire byĀ movingĀ it in a magnetic field, or by moving the magnet or, if the magnetic field is caused by some other current, by changing the current in that wire. It’s referred to as the ‘flux rule’, or Faraday’s Law. Remember: we’ve seen Gauss’ Law, then AmpĆØre’s Law, and then that Biot-Savart Law, and so now it’s time for Faraday’s Law. šŸ™‚Ā Faraday’s Law is Maxwell’s third equation really, aka as the Maxwell-FaradayĀ Law of Induction:

āˆ‡Ć—E =Ā āˆ’āˆ‚B/āˆ‚t

Now you’ll wonder: what’sĀ flux got to do with this formula? āˆ‡Ć—E is aboutĀ circulation, not about flux!Ā Well… Let me copy Feynman’s answer:

Faraday's law

So… There you go. And, yes, you’re right, instead of writing Faraday’s Law asĀ āˆ‡Ć—E =Ā āˆ’āˆ‚B/āˆ‚t, we should write it as:

emf

That’s a easier to understand, and it’s also easier to work with, as we’ll see in a moment. So the point is: whenever the magnetic fluxĀ changes, there’s a push on the electrons in the wire. That push is referred to as theĀ electromotive force, abbreviated as emfĀ or EMF, and so it’s that line and/or surface integral above indeed. Let me paraphrase FeynmanĀ so you fullyĀ understand what we’re talking about here:

When we move our wire in a magnetic field, or when we move a magnet near the wire, or when we change the current in a nearby wire, there will be some net push on the electrons in the wire in one direction along the wire. There may be pushes in different directions at different places, but there will be more push in one direction than another. What counts is the push integrated around the complete circuit. We call this net integrated push the electromotive force (abbreviated emf) in the circuit. More precisely, the emf is defined as the tangential force per unit charge in the wire integrated over length, once around the complete circuit.

So that’s the integral. šŸ™‚ And that’s how we can turn that motor above into a generator:Ā instead of putting a current through the wire to make it turn, we can turn the loop, by hand or by a waterwheel or by whatever. Now, when the coil rotates, its wires will be moving in the magnetic field and so we will find an emf in the circuit of the coil, and so that’s how the motor becomes a generator.

Now, let me quickly interject something here: when I say ‘a push on the electrons in the wire’, what electrons are we talking about? How many? Well… I’ll answer that question in very much detail in a moment but, as for now, just note that the emfĀ is some quantity expressedĀ per coulombĀ or, as Feynman puts it above, per unit charge.Ā So we’ll need to multiply it with the current in the circuit to get theĀ powerĀ of our little generator.

OK. Let’s move on. Indeed, all I can doĀ here is mention just a few basics, so we can move on to the next thing. If you really want to know all of the nitty-gritty, then you should just readĀ Feynman’sĀ LectureĀ on induced currents. That’s got everything. And, no, don’t worry: contrary to what you might expect, my ‘basics’ do not amount to a terrible pile of formulas. In fact, it’s all easy and quite amusing stuff, and I should probably include a lot more. But then… Well… I always need to move on… If not, I’ll never get to the stuff that I reallyĀ want to understand. 😦

The electromotive force

We defined the electromotive force above, including its formula:

emf

What are the units? Let’s see… We know B was measuredĀ notĀ in newton per coulomb, like the electric field E, but in NĀ·s/CĀ·m, because we had to multiply the magnetic field strength with the velocity of the charge to find the force per unit charge, cf. theĀ F/q = vƗBĀ equation. Now what’s the unit in which we’d express that surface integral? We must multiply with m2, so we getĀ NĀ·mĀ·s/C. Now let’s simplify that by noting that one volt is equal to 1 NĀ·m/C. [The voltĀ has a number of definitions, but the one that applies here is that it’s the potential difference between two points that will impart one jouleĀ (i.e. 1 NĀ·m) of energyĀ to a unit of charge (i.e. 1 C) that passes between them.] So we can measure the magnetic flux in volt-seconds, i.e. VĀ·s.Ā And then we take the derivative in regard to time, so we divide by s, and so we get… Volt! The emf is measured in volt!

Does that make sense? I guess so: the emf causes a current, just like a potential difference, i.e. a voltage, and, therefore, we can and should look at the emf as a voltage too!

But let’s think about it some more, though. In differential form,Ā Faraday’s Law, is just that āˆ‡Ć—E =Ā āˆ’āˆ‚B/āˆ‚t equation, so that’s just one of Maxwell’s four equations, and so we prefer to write it as the “flux rule”. Now, the “flux rule” says that the electromotive force (abbreviated as emf or EMF) on the electrons inĀ a closed circuit is equal to the time rate of change of the magnetic flux it encloses. As mentioned above, we measureĀ magneticĀ flux inĀ volt-secondsĀ (i.e. VĀ·s), so its time rate of change is measured in volt (because the time rate of change is a quantity expressed per second), and so the emfĀ is measured in volt, i.e. jouleĀ perĀ coulomb, as 1 V = 1 NĀ·m/CĀ = 1 J/C.Ā What does it mean?

The time rate of change of the magnetic flux can changeĀ because the surface covered by our loop changes,Ā orĀ because the field itself changes, or by both. Whatever the cause, it will change the emf, or the voltage, and so it will make the electrons move. So let’s suppose we have some generatorĀ generating someĀ emf.Ā The emfĀ can be used to do someĀ work. We can charge a capacitor, for example. So how would that work?

More charge on the capacitor will increase the voltage V of the capacitor, i.e. the potential difference V = Φ1 āˆ’ Φ2 between the two plates. Now, we know that the increase of the voltage V will be proportional to the increase of the charge Q, and that the constant of proportionality is definedĀ by the capacity C of the capacitor: C = Q/V. [How do we know that? Well… Have a look at my post on capacitors.] Now, if our capacitor has an enormous capacity, then its voltage won’t increase very rapidly. However, it’s clear that, no matter how large the capacity, its voltage will increase. It’s just a matter of time. Now, its voltage cannot be higher than the emfĀ provided by our ‘generator’, because it will then want to discharge through the same circuit!

So we’re talking power and energy here, and so we need to put some loadĀ on our generator. Power is the rate of doing work, so it’s the time rate of change of energy, and it’s expressed in joule per second. The energy of our capacitor isĀ U =Ā (1/2)Ā·Q2/C =Ā (1/2)Ā·CĀ·V2.Ā [How do we know that? Well… Have a look at my post on capacitorsĀ once again. :-)] So let’s take the time derivative of U assumingĀ some constant voltageĀ V. We get: dU/dt = d[(1/2)Ā·Q2/C]/dt = (Q/C)Ā·dQ/dt = VĀ·dQ/dt. So that’s the powerĀ that the generator would need to supply to charge the generator. As I’ll show in a moment, the power supplied by a generator is, indeed, equal to theĀ emf times theĀ current, and the current is the time rate of change of the charge, so I = dQ/dt.

So, yes, it all works out: the power that’s being supplied by our generator will be used to charge our capacitor. Now, you may wonder: what about the current? Where is the current in Faraday’s Law?Ā The answer is: Faraday’s Law doesn’t have the current. It’s just not there. The emf is expressed in volt, and so that’s energy per coulomb, so it’s per unit charge. How much power an generator can and will deliver depends on its design, and the circuit and load that we will be putting on it. So we can’t say howĀ many coulomb we will have. It all depends. But you can imagine that, if the loop would be bigger, or if we’d have a coil with many loops, then our generator would be able to produce more power, i.e. it would be able to move more electrons, so the mentioned power = (emf)Ɨ(current) product would be larger. šŸ™‚

Finally, to conclude, note Feynman’s definition of the emf: the tangential force per unit charge in the wire integrated over length around the complete circuit. So we’ve got force times distance here, but per unit charge. Now, force times distance is work, or energy, and so… Yes, emfĀ is joule per coulomb, definitely! šŸ™‚

[…] Don’t worry too much if you don’t quite ‘get’ this. I’ll come back to it when discussingĀ electric circuits, which I’ll do in my next posts.

Self-inductance andĀ Lenz’s rule

We talked about motors and generators above. We also haveĀ transformers, like the one below. What’s going on here is that anĀ alternating currentĀ (AC) produces a continuously varying magnetic field, which generates an alternating emf in the second coil, which produces enough power to light an electric bulb.

transformer

Now, the total emf in coil (b) is the sum of the emf’s of the separate turns of coil, so if we wind (b) with many turns, we’ll get a larger emf, so we can ‘transform’ the voltage to some other voltage. From your high-school classes, you should know howĀ thatĀ works.

The thing I want to talk about here is something else, though. There is an induction effect in coil (a) itself. Indeed, the varying current in coilĀ (a) produces a varying magnetic field inside itself, and the flux of this field is continually changing, so there is a self-induced emf in coilĀ (a). The effect is called self-inductance, and so it’s the emf acting on a current itself when it is building up a magnetic field or, in general, when its field is changing in any way.Ā It’s a most remarkable phenomenon, and so let me paraphrase Feynman as he describes it:

“When we gave ā€œthe flux ruleā€ that the emf is equal to the rate of change of the flux linkage, we didn’t specify the direction of the emf. There is a simple rule, called Lenz’s rule, for figuring out which way the emf goes: the emf tries to oppose any flux change. That is, the direction of an induced emf is always such that if a current were to flow in the direction of the emf, it would produce a flux of B that opposes the change inĀ BĀ that produces the emf.Ā In particular, if there is a changing current in a single coil (or in any wire), there is a ā€œbackā€ emf in the circuit. This emf acts on the charges flowing in the coil to oppose the change in magnetic field, and so in the direction to oppose the change in current. It tries to keep the current constant; it is opposite to the current when the current is increasing, and it is in the direction of the current when it is decreasing. A current in a self-inductance has ā€œinertia,ā€ because the inductive effects try to keep the flow constant, just as mechanical inertia tries to keep the velocity of an object constant.”

Hmm… That’s something you need to read a couple of times to fully digest it. There’s a nice demo on YouTube, showing an MIT physics videoĀ demonstrating this effect with a metal ring placed on the end of an electromagnet. You’ve probably seen it before: the electromagnet is connected to a current, and the ring flies into the air. The explanation is that theĀ induced currentsĀ in the ring create a magnetic fieldĀ opposing the change of field through it. So the ring and the coil repel just like two magnets with opposite poles. The effect is no longer there when a thin radial cut is made in the ring, because then there can be no current. The nice thing about the video is that it shows how the effect gets much more dramatic when anĀ alternatingĀ current is applied, rather than a DC current. And it also shows what happens when you first cool the ring in liquid nitrogen. šŸ™‚

You may also notice the sparks when the electromagnet is being turned on. Believe it or not, that’s also related to a “back emf”. Indeed, when we disconnect a large electromagnet by opening a switch, the current is supposed to immediately go to zero but, in trying to do so, it generates a large “back emf”: large enough to develop an arc across the opening contacts of the switch. The high voltage is also not good for the insulation of the coil, as it might damage it. So that’s why large electromagnets usually include some extra circuit, which allows the “back current” to discharge less dramatically. But I’ll refer you to Feynman for more details, as any illustration here would clutter the exposĆ©.

Eddy currents

I like educational videos, and so I should give you a few references here, but there’s so many of this that I’ll let youĀ googleĀ a few yourself. The most spectacular demonstration of eddy currents is those that appear in a superconductor: even back in the 1970s, when Feynman wrote hisĀ Lectures, the effect ofĀ magnetic levitationĀ was well known. Feynman illustrates the effect with the simple diagram below: when bringing a magnet near to a perfect conductor, such as tinĀ below 3.8°K, eddy currents will create opposing fields, so that no magnetic flux entersĀ the superconducting material. The effect is also referred to as theĀ MeisnerĀ effect, after the German physicist Walther Meisner, although it was discovered much earlier (in 1911) by a Dutch physicist in Leiden, Heike Kamerlingh Onnes, who got a Nobel Prize for it.

superconductor

Of course, we have eddy currents in less dramatic situations as well. The phenomenon of eddy currents is usually demonstrated by the braking of a sheet of metal as it swings back and forth between the poles of an electromagnet, as illustrated below (left). The illustration on the right shows how eddy-current effect can be drastically reduced by cutting slots in the plate, so that’s like making a radial cut in our jumping ring. šŸ™‚

eddy currentseddy currents 2

The Faraday disc

The Faraday disc is interesting, not only from a historical point of view – the illustration below is a 19th century model, so Michael Faraday may have used himself – but also because it seems to contradict the “flux of rule”: as the disc rotates through aĀ steadyĀ magnetic field, it will produce some emf, but so there’s no change in the flux. How is that possible?

Faraday_disk_generatorFaraday disk

The answer, of course, is that we are ‘cheating’ here: the material is moving, so we’re actually moving the ‘wire’, or the circuit if you want, so here we need to combine twoĀ equations:

two laws

If we do that, you’ll see it all makes sense. šŸ™‚ Oh… That Faraday disc is referred to as a homopolar generator, and it’s quite interesting. You should check out what happened to the concept in the Wikipedia article on it. The Faraday disc was apparently used as a source for power pulsesĀ in the 1950s. The thing below could store 500Ā mega-joules and deliver currents up to 2Ā mega-ampĆØre, i.e. 2 millionĀ amps!Ā Fascinating, isn’t it? šŸ™‚800px-Homopolar_anu-MJC

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/

Magnetic dipoles and their torque and energy

We studied the magnetic dipole in very much detail in one of my previous posts but, while we talked about an awful lot of stuff there, we actually managed toĀ notĀ talk about theĀ torqueĀ on a it, when it’s placed in the magnetic field ofĀ otherĀ currents, or some other magnetic field tout court. Now, that’s what drives electric motors and generators, of course, and so weĀ shouldĀ talk about it, which is what I’ll do here. Let me first remind you of the concept of torque, and then we’ll apply it to a loop of current. šŸ™‚

The concept of torque

The concept of torque is easy to grasp intuitively, but the math involved isĀ notĀ so easy. Let me sum up the basics (for the detail, I’ll refer you to my posts on spin and angular momentum). In essence, for rotations in space (i.e. rotational motion), the torque is what the force is for linear motion:

  1. It’s the torque (Ļ„) that makes an object spin faster or slower around some axis, just like the force would accelerate or decelerate that very same object when it would be moving along some curve.
  2. There’s also a similar ā€˜law of Newton’ for torque: you’ll remember that the force equals the time rate of change of a vector quantity referred to as (linear) momentum: F = dp/dt = d(mv)/dt = ma (the mass times the acceleration). Likewise, we have a vector quantity that is referred to as angular momentum (L), and we can write: Ļ„ (i.e. the Greek tau) = dL/dt.
  3. Finally, instead of linear velocity, we’ll have an angular velocity ω (omega), which is the time rate of change of the angle Īø that defines how far the object has goneĀ around (as opposed to the distance in linear dynamics, describing how far the object has gone along). So we have ω = dĪø/dt. This is actually easy to visualize because we know that Īø, expressed in radians, is actually the length of the corresponding arc on the unit circle. Hence, the equivalence with the linear distance traveled is easily ascertained.

There are many moreĀ similarities, like an angular acceleration: α = dω/dt = d2Īø/dt2, and we should also note that, just like the force, the torque is doing work – in its conventional definition asĀ used in physics – as it turns an object instead of justĀ moving it, so we can write:

Ī”W = τ·Δθ

SoĀ it’s all the same-same but different once more šŸ™‚ and so now we also need to point out some differences. The animation below does that very well, as it relates the ā€˜new’ concepts – i.e. torque and angular momentum – to the ā€˜old’ concepts – i.e. force and linear momentum. It does so using theĀ vector cross product, which is really all you need to understand the math involved. Just look carefully at all of the vectors involved, which you can identify by their colors, i.e. red-brownĀ (r),Ā light-blueĀ (Ļ„), dark-blueĀ (F), light-green (L), and dark-greenĀ (p).

Torque_animation

So what do we have here? We have vector quantities once again, denoted by symbols inĀ bold-face. Having said that, I should note thatĀ Ļ„, L and ω are ‘special’ vectors: they are referred to asĀ axial vectors, as opposed to the polar vectors F, p and v. To put it simply: polar vectors represent somethingĀ physical, and axial vectors are more likeĀ mathematicalĀ vectors, but that’s a very imprecise and, hence, essential non-correct definition. šŸ™‚Ā Axial vectors are directed along theĀ axisĀ of spin – so that is, strangely enough, at right angles to the direction of spin, or perpendicular to the ā€˜plane of the twist’ as Feynman calls it – and the direction of the axial vector is determined by a convention which is referred to as the ā€˜right-hand screw rule’. šŸ™‚

Now, I know it’s not so easy to visualize vector cross products, so it may help to first think of torque (also known, for some obscure reason, as the moment of the force) as a twist on an object or a plane. Indeed, the torque’s magnitudeĀ can be defined in another way: it’s equal to the tangential component of the force, i.e. FĀ·sin(Δθ), times the distance between the object and the axis of rotation (we’ll denote this distance by r). This quantity is also equal to the product of the magnitude of the force itself and the length of the so-called lever arm, i.e. the perpendicular distance from the axis to the line of action of the force (this lever arm length is denoted by r0). So, we can define Ļ„ without the use of the vector cross-product, and in not less than three different ways actually. Indeed, the torque is equal to:

  1. The product of the tangential component of the force times the distance r: Ļ„ = rĀ·Ft= rĀ·FĀ·sin(Δθ);
  2. The product of the length of the lever arm times the force: Ļ„ = r0Ā·F;
  3. The work done per unit of distance traveled: Ļ„ = Ī”W/Δθ or Ļ„ = dW/dĪø in the limit.

Phew!Ā Yeah. I know. It’s not so easy… However, I regret to have to inform you that you’ll need to go evenĀ further in your understanding of torque. More specifically, you really need to understand why and how we define the torque as a vector cross product, and so please do check out that post of mine on the fundamentals of ‘torque math’. If you don’t want to do that, then just try to remember the definition of torque as an axial vector, which is:

τ = (τyz, τzx, τxy) = (τx, τy, τz) with

Ļ„x = Ļ„yz = yFz – zFyĀ (i.e. the torque about the x-axis, i.e. in the yz-plane),

Ļ„y = Ļ„zx = zFx – xFzĀ (i.e. the torque about the y-axis, i.e. in the zx-plane), and

Ļ„z = Ļ„xy = xFy – yFxĀ (i.e. the torque about the z-axis, i.e. in the xy-plane).

The angular momentum L is defined in the same way:

L = (Lyz, Lzx, Lxy) = (Lx, Ly, Lz) with

Lx = Lyz = ypz – zpy (i.e. the angular momentum about the x-axis),

Ly = Lzx = zpx – xpz (i.e. the angular momentum about the y-axis), and

Lz = Lxy = xpy – ypx (i.e. the angular momentum about the z-axis).

Let’s nowĀ apply the concepts to a loop of current.

The forces on a current loop

The geometry of the situation is depicted below. I know it looks messy but let me help you identifying the moving parts, so to speak. šŸ™‚ We’ve got a loop with current and so we’ve got a magnetic dipole with some moment μ. From my post on the magnetic dipole, you know that μ‘s magnitude is equal to |μ| = μ = (current)Ā·(area of the loop) = IĀ·aĀ·b.

Geometry 2

Now look at the B vectors, i.e. the magnetic field.Ā Please note that these vectors represent some external magnetic field!Ā So it’sĀ notĀ like what we did in our post on the dipole: we’re not looking at the magnetic fieldĀ causedĀ by our loop, but at how it behaves in some externalĀ magnetic field. Now, because it’s kinda convenient to analyze, we assume that the direction of our external field B is the direction of the z-axis, so that’s what you see in this illustration: the B vectors all point north.Ā Now look at the force vectors, remembering that the magnetic force is equal to:

FmagneticĀ = qvƗB

So that gives the F1,Ā F2,Ā F3, andĀ F4Ā vectors (so that’s the force on the first, second, third and fourth leg of the loop respectively) the magnitude and direction they’re having. Now, it’s easy to see that the opposite forces, i.e. theĀ F1–F2Ā andĀ F3–F4Ā pair respectively, create aĀ torque. The torque because ofĀ F1Ā andĀ F2Ā is a torque which will tend to rotate the loopĀ about the y-axis, so that’s a torque in the xz-plane, while the torque because ofĀ F3Ā andĀ F4Ā will be some torque about the x-axis and/or the z-axis. As you can see, the torque is such that it will try to line up the moment vector μ with the magnetic field B. In fact, the geometry of the situation above is such thatĀ F3Ā andĀ F4Ā have already done their job, so to speak: the moment vector μ is already lined up with the xz-plane, so there’s not net torque in that plane. However, that’s just because of the specifics of the situation here: the more general situation is that we’d have some torque about all three axes, and so we need to find that vectorĀ Ļ„.

If we’d be talking someĀ electricĀ dipole, the analysis would be veryĀ straightforward, because the electric force is justĀ FelectricĀ = qE, which we can also write as E = FelectricĀ =/q, so the field is just the force on oneĀ unit of electricĀ charge, and so it’s (relatively) easy to see that we’d get the following formula for the torque vector:

Ļ„ = pƗE

Of course, the p is the electric dipole momentĀ here, not someĀ linear momentum. [And, yes, please do try to check this formula. Sorry I can’t elaborate on it, but the objective of this blog is notĀ substitute for a textbook!]

Now, all of the analogies between the electric and magnetic dipole field, which we explored in the above-mentioned post of mine, would tend to make us think that we can write Ļ„ here as:

Ļ„ = μ×B

Well… Yes. It works. Now you may want to know why it works šŸ™‚ and so let meĀ give you the following hint. Each charge in a wire feels thatĀ FmagneticĀ = qvƗB force, so the total magnetic force on some volume Ī”V, which I’ll denote by Ī”F for a while, is the sum of the forces on all of the individual charges. So let’s assume we’ve got N charges per unit volume, then we’ve got NĀ·Ī”V charges in our little volume Ī”V, so we write:Ā Ī”FĀ = NĀ·Ī”VĀ·qĀ·vƗB. You’re probably confused now: what’s the v here? It’s the (drift) velocity of the (free) electrons that make up our current I. Indeed, the protons don’t move. šŸ™‚ So NĀ·qĀ·v is just the current density j, so we get: Ī”FĀ = jƗBĪ”V, which implies that the forceĀ per unit volumeĀ is equal to jƗB. But we need to relate it to theĀ currentĀ in our wire, not the current density. Relax. We’re almost there. TheĀ Ī”V in a wire is just its cross-sectional area A times some length, which I’ll denote byĀ Ī”L, so Ī”FĀ = jƗBĪ”V becomes Ī”FĀ = jƗBAĪ”L. Now, jA is the vector current I, so we get the simple result we need here: Ī”FĀ = IƗBĪ”L, i.e. Ā the magnetic force per unit length on a wire is equal toĀ Ī”F/Ī”L = IƗB.

Let’s now get back to our magnetic dipole and calculate F1Ā andĀ F2. The length of ‘wire’ is the length of the leg of the loop, i.e. b, so we can write:

F1Ā =Ā āˆ’F2Ā = bĀ·IƗB

So the magnitude of these forces is equal F1Ā =Ā F2Ā = IĀ·BĀ·b. Now,Ā The length of the moment orĀ lever armĀ is, obviously, equal to aĀ·sinĪø, so the magnitude of the torque is equal to the force times the lever arm (cf. theĀ Ļ„ = r0Ā·F formula above) and so we can write:

Ļ„ =Ā IĀ·BĀ·bĀ·aĀ·sinĪø

But I·a·b is the magnitude of the magnetic moment μ, so we get:

Ļ„ = μ·BĀ·sinĪø

Now that’s consistent with the definition of the vector cross product:

Ļ„ = μ×BĀ = |μ|Ā·|B|Ā·sinĪøĀ·n = μ·BĀ·sinĪøĀ·n

Done! Now,Ā electric motors and generators are all aboutĀ workĀ and, therefore, we also need to briefly talk about energy here.

The energy of a magnetic dipole

Let me remind you that we could also write the torque as the work done per unit of distance traveled, i.e. as Ļ„ = Ī”W/Δθ or Ļ„ = dW/dĪø in the limit. Now, the torque tries to line up the moment with the field, and so the energy will beĀ lowestĀ when μ and B are parallel, so we need to throw in a minus sign when writing:

Ā Ļ„ = āˆ’dU/dĪø ⇔ dU =Ā āˆ’Ļ„Ā·dĪø

We should now integrate over the [0, Īø] interval toĀ find U, also using our Ļ„ = μ·BĀ·sinĪø formula. That’s easy, because we know that d(cosĪø)/dĪø =Ā āˆ’sinĪø, so that integral yields:

U = 1 āˆ’ μ·BĀ·cosĪø + a constant

If we choose the constant to be zero, and if we equate μ·B with 1, we get the blue graph below:

graph energy magnetic dipole 3

The μ·B in the U = 1 āˆ’ μ·BĀ·cosĪø formula is just a scaling factor, obviously, so it determines the minimumĀ andĀ maximumĀ energy. Now, you may want to limitĀ the relevant range of Īø to [0, π], but that’s not necessary: the energy of our loop of current does go up and down as shown in the graph. Just think about it: it all makes perfect sense!

Now, there is, of course, more energy in the loop than this U energy because energy is needed to maintain the current in the loop, and so we didn’t talk about that here. Therefore, we’ll qualify this ‘energy’ and call it the mechanicalĀ energy, which we’ll abbreviate byĀ Umech. In addition, we could, and will, choose some other constant of integration, so that amounts to choosing some other reference point for the lowest energy level. Why? Because it then allows us to write UmechĀ as a vectorĀ dotĀ product, so we get:

Umech =Ā āˆ’Ī¼Ā·BĀ·cosĪø = āˆ’Ī¼Ā·B

The graph is pretty much the same, but it now goes from āˆ’Ī¼Ā·B to +μ·B, as shown by the red graph in the illustration above.

Finally, you should note that the Umech = āˆ’Ī¼Ā·BĀ formula is similar to what you’ll usually see written for the energy of an electric dipole:Ā UĀ =Ā āˆ’pĀ·E.Ā So that’s all nice and good! However, you should remember that the electrostatic energy of anĀ electric dipole (i.e. two opposite charges separated by some distance d) isĀ all of the energy, as we don’t need to maintain some current to create the dipole moment!

Now, Feynman does all kinds of things with these formulas in his LecturesĀ on electromagnetism but I really think this is all you need to know about it—for the moment, at least. šŸ™‚

The magnetic field of circuits: the Law of Biot and Savart

Pre-script (dated 26 June 2020): This post got mutilated by the removal of some material by the dark force. You should be able to follow the main story line, however. If anything, the lack of illustrations might actually help you to think things through for yourself. In any case, we now have different views on these concepts as part of our realist interpretation of quantum mechanics, so we recommend you read our recent papers instead of these old blog posts.

Original post:

We studied the magnetic dipole in very much detail in one of my previous posts. While we talked about an awful lot of stuff there, we actually managed toĀ notĀ talk about theĀ torqueĀ on a it, when it’s placed in the magnetic field ofĀ otherĀ currents. Now, that’s what drives electric motors and generators, of course, and so weĀ shouldĀ talk about it, which is what I’ll do in my next post. Before doing so, however, I need to give you one or two extra formulas generalizing some of the results we obtained in our previous posts on magnetostatics. So that’s what I do under this heading: the magnetic field of circuits.Ā The idea is simple: loops of current are not always nice squares or circles. Their shape might be quite irregular, indeed, like the loop below.

irregular loop

Of course, the same general formula should apply. So we can find the magnetic vector potential with the following integral:

loop of current V

Just to make sure, let me re-insert its equivalent forĀ electrostatics, so you can see they’reĀ (almost) the same:

formula 1

But we’re talking a wire here, so how can we relate the current densityĀ j and the volume elementĀ dV to that? It’s easy: the illustration below shows that we can simply write:

jĀ·dV = jĀ·SĀ·ds = IĀ·ds

volume and wire

Therefore, we can write our integral for the vector potential as:

loop of current I

Of course, you should note the subtle change from aĀ volumeĀ integral to aĀ lineĀ integral, so it’s not all that straightforward, but we’re good to go.Ā Now, in electrostatics, we actually had a fairly simple integral for the electric field itself:

formula for E no 2

To be clear, E(1) is the field of a known charge distribution, which is represented by ρ(2), at pointĀ (1). The integral isĀ almost the same as the one for Φ, but we’re talkingĀ vectorsĀ hereĀ (E andĀ e12) rather than scalarsĀ (ρ and Φ), and you should also note theĀ squareĀ in the denominator of the integral. šŸ™‚

As you might expect, there is a similar integral for B, which we find by… Well… We just need to calculate B, so that’s the curl of A:

formula for B integral

How do we do that? It’s not so easy, so let me just copy the master himself:

integral calculation

So this integral gives B directly in terms of the known currents. The geometry involved is easy but, just in case, Feynman illustrates it, quite simply, as follows:

geometry

Now, there’s one more step to take, and then we’re done. If we’re talking a circuit of small wire, then we can replaceĀ jĀ·dV by IĀ·dsĀ once more, and, hence, we get the Biot-Savart Law in its final form:

B formula 2

Note the minus sign: it appears because we reversed the order of the vector cross product, and also note we actually have threeĀ integrals here, one for each componentĀ of B, so that’sĀ just like that integral for A.

So… That’s it. šŸ™‚ I’ll conclude by two small remarks:

  1. The law is named after Jean-Baptiste Biot and FĆ©lix Savart, two incredible Frenchmen (it’s really worthwhile checking their biographies on Wikipedia), who jotted it down inĀ 1820, so that’s almost 200 years ago. Isn’t that amazing?
  2. You see we sort of got rid of the vector potential with this formula. So the question is: ā€œWhat is the advantage of the vector potential if we can findĀ BĀ directly with a vector integral? After all, AĀ also involves three integrals!ā€ I’ll let Feynman reply to that question:

Because of the cross product, the integrals for BĀ are usually more complicated. Also, since the integrals forĀ AĀ are like those of electrostatics, we may already know them. Finally, we will see that in more advanced theoretical matters (in relativity, in advanced formulations of the laws of mechanics, like the principle of least action to be discussed later, and in quantum mechanics), the vector potential plays an important role.

In fact, Feynman makes the point on the vector potential being relevant veryĀ explicit by just boldly stating two laws in quantum mechanics in which the magnetic and electric potential are used, not the magnetic or electric field. Indeed, it seems an external magnetic or electric fieldĀ changes probability amplitudes. I’ll just jot down the two laws below, but leave it to you to decide whether or not you want to read the whole argument.Ā qm1qm2

The key point that Feynman is making is that Φ and A are equally ‘real’ or ‘unreal’ as E and B in terms of explaining physical realities. I get the point, but I don’t find it necessary to copy the whole argument here. Perhaps it’s sufficient to just quote Feynman’s introduction to it, which says it all, in my humble opinion, that is:

“There are many changes in what concepts are important when we go from classical to quantum mechanics. We have already discussed some of them in Volume I. In particular, the force concept gradually fades away, while the concepts of energy and momentum become of paramount importance. You remember that instead of particle motions, one deals with probability amplitudes which vary in space and time. In these amplitudes there are wavelengths related to momenta, and frequencies related to energies. The momenta and energies, which determine the phases of wave functions, are therefore the important quantities in quantum mechanics. Instead of forces, we deal with the way interactions change the wavelength of the waves. The idea of a force becomes quite secondary—if it is there at all. When people talk about nuclear forces, for example, what they usually analyze and work with are the energies of interaction of two nucleons, and not the force between them. Nobody ever differentiates the energy to find out what the force looks like. In this section we want to describe how the vector and scalar potentials enter into quantum mechanics. It is, in fact, just because momentum and energy play a central role in quantum mechanics that A and Φ provide the most direct way of introducing electromagnetic effects into quantum descriptions.”

OK. That’s sufficient really.Ā Onwards!

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/

Gravitational waves: how should we imagine them?

This post is not a post. It’s just a reminder for myself to look into gravitational wavesĀ at some point in time. We know how electromagnetic waves travel through space: they do so because of the mechanism described in Maxwell’s equation: a changing magnetic field causes a changing electric field, and a changing magnetic field causes a (changing) electric field, as illustrated below.

Maxwell interaction

Electromagneticwave3D

So… Electromagnetism isĀ one phenomenon only, but we do analyze the E and BĀ fields as separate things, as the equations below illustrate:

formula for E

magnetic field

B is co-defined with E, and then all of the dynamics work themselves out through theĀ āˆ‚E/āˆ‚t and āˆ‚E/āˆ‚t functions. Now, when talking gravity, we only haveĀ positiveĀ ‘charges’, referred to as masses, alwaysĀ attracting each other, but that doesn’t matter all that much: Coulomb’s and Newton’s Law have the same mathematical structure (as shown below), except for the minus sign, and so thereĀ mustĀ be someĀ equivalent to the electromagnetic wave, explaining gravitational waves , using the same mathematical concepts, as aĀ propagationĀ of a force on a ‘unit charge’ (so that’s a unit massĀ in this case) using the very same concepts, i.e. the concepts of a field – twoĀ separate fields, I should say, just like E and BĀ  – and of its fluxĀ and circulation.Ā gravity law

550px-CoulombsLaw

So we’d have an EGĀ and a BGĀ field, so to speak, or a GEĀ and a GBĀ field, and formulas for the flux and circulation of both, resembling Maxwell’s equations. In fact, they’d be the same, more or less. It’s a powerful idea, and I am sure the idea has been fully developed elsewhere. In fact, I quicklyĀ googledĀ a few things, but it seems that the whole area is a pretty new area of research, both theoretically as well as experimentally—to my surprise!

Hmm… Interesting idea, but I’ll need to do a lot more analysis to be able to grind through this… One day, I will… :-/ The first thing I need to do, obviously, is toĀ thoroughlyĀ study how these equations for E and B above can be derived from Maxwell’s equations. I’ll need some time for that, and then I can see if it’s consistent with Einstein’s special andĀ generalĀ theories of relativity.

I’ll update you on progress. šŸ™‚

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 20, 2020 as a result of a DMCA takedown notice from Michael A. Gottlieb, Rudolf Pfeiffer, and The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/

The field of electric and magnetic dipoles

You’l surely remember theĀ electricĀ field of an electric dipole, as depicted below:dipole field 2

We distinguish two components:

  1. A z-component along the axis of the dipole itself, and
  2. A so-calledĀ transverseĀ component, i.e. the component that isĀ perpendicularĀ to the axis of the dipole, or the z-axis.

Pythagoras’ rule then gives us theĀ totalĀ field, which is equal to:

components electric field

I’ll give you the formulas for both components in a moment, but let’s first introduce the concept of a magneticĀ dipole. Look atĀ the magneticĀ field of a solenoid below, and imagine we reduce the solenoid to one loop of current only. What would we get?

graph solenoid

We get aĀ magneticĀ field that resembles the electric field from an electric dipole. Of course, it’s aĀ magneticĀ field, and it’s not the field of an electric dipole but of a magnetic dipole which, in this case, is the field of a small loop of current. Feynman depicts the geometry of the situation, and the relevant variables, as follows:

diagram magnetic dipole

Now, in my previous post, in which I presentedĀ the magnetic vector potential, I pointed out that the equations for the x-, y- and z-component of the vector potential A are Poisson equations and, therefore, they are mathematically identical to the Poisson equation for electrostatics. Now that’s why the x-, y- and z-component of the vector potentialĀ arising from a current density j is exactly the same as the electric potential Φ that would be produced by a charge density ρ equal to jx, jyĀ  and jzĀ respectively, divided by c2, and we use that fact to calculate them.Ā Huh?Ā Yes.Ā Let me just jot it down the equations once more, so you can appreciate what I am saying here:

āˆ‡2Φ = –ρ/ε0Ā (Poisson equation for electrostatics)

āˆ‡2A = –j/ε0c2Ā (Poisson equation for magnetostatics)

⇔ AxĀ = –jx/ε0c2, AyĀ = –jy/ε0c2Ā and AzĀ = –jz/ε0c2

We didn’t go through all of the nitty-gritty of the exercise for a straight wire and a solenoid in our previous post, but we will do so now for our single loop of current. It’s kinda tedious stuff, so just hang in there and try to get through it. šŸ™‚

We’ll start with Ax. To calculateĀ Ax, we need jx, so that’s the component of the current density that’s shown below. [As you can see, jz is zero, and we’ll talk about jyĀ in a minute.]

diagram magnetic dipole 2What are those + and āˆ’ signs next to those two loop legs? Well… Think about what we wrote above: we need to think of jxĀ divided by c2Ā as some charge density, and the associated Φ will then equalĀ Ax. It’s an easy principle:Ā the same equations have the same solutions. We’re only switching symbols. So look at the geometry of the situation: we have a uniformĀ current and, hence, a uniform current density, in each leg of the loop, so jxĀ is some constant over the two legs shown here. Therefore, theĀ electrostatic equivalent is that of two charged rods with some charge density Ī».

Now, it’s easy to see that λ =Ā jx/c2Ā = I/c2Ā and… So… Well… Hmm… […] So how do we calculate the electrostatic potential Φ from two charged rods? Well… We haven’t calculated that in any of our posts, but if R is ‘large enough’Ā (and it should be, because we’re interested in the fields only at distances that are large compared to the size of the loop), then Φ will be the potential of a dipole with (opposite) charges that are equal to the (opposite) charges of the rods. Now, Ī» is a charge density, so we need to multiply it with the length of the rods a to get the charge. And then we need to multiply the charge with the distance between the two charges to get the dipole moment, so we write:

p = λab = Iab/c2

OK. That’s a nice result because we can now use that to calculate the potential. Indeed, you may or may not remember the formula for the potential of an electric dipole, but we wrote it as follows inĀ one of my posts:

Φ = āˆ’pĀ·āˆ‡Ļ†0Ā = āˆ’(1/4πε0)pĀ·āˆ‡(1/R) =Ā āˆ’(1/4πε0)(pĀ·er)/R2

āˆ‡Ļ†0 is theĀ gradientĀ of φ0, and φ0Ā is the potential of a unitĀ point of charge: φ0Ā = 1/4πε0r. In any case, don’t worry too much about this formula right now, because I’ll give you some easier version of it later on. So let’s just move on. To calculate that vector productĀ pĀ·erĀ = |p|Ā·|erĀ |Ā·cosĪø = pĀ·cosĪø, we need to know the angle Īø between them, and the direction of the dipole moment. That’s simple enough: the vectorĀ p points in theĀ negativeĀ y-direction, so cosĪø = –y/R, where y is a coordinate of P. So we have:

formula 4

Now we do the grand substitution, the hat-trick really: Φ becomes Ax, and λ = jx/c2 = I/c2, and so we get:

formula 5

Huh?Ā Yes. It’s a great trick really. Brilliant! šŸ™‚ You’ll get it. Just think about what happened for a while.

Next thing is jy. So what about it? Well… As you may have suspected, it’s a perfectly symmetrical situation, so we can just repeat the same reasoning but swap x for y and y for x. We get:

formula 7

Note the minus sign has disappeared, but you can explain that yourself. And, of course, we only have the vector potential here, so we still need to calculate B from it, using theĀ B =Ā āˆ‡Ć—A equation, which is three equations involving all possible cross-derivatives really:

formula for BPhew!Ā It looks quite horrible indeed!Ā Feynman says: “It’s a little more complicated than electrostatics, but the same idea.” But… Well… I guess it’s a bit hard to notĀ take this statement with a pinch of salt.Ā For starters, there is no intuitive physical interpretation of the magnetic vector potentialĀ A, as opposed to the electric potential Φ. More importantly, calculatingĀ E = āˆ’āˆ‡Ī¦ involves only three derivatives, and that’s a helluva lot easier than calculating not less than six cross-derivatives and keep track of their order and all of that. In any case, nothing much we can do about, so let’s grind through it. :-/ And to be fully honest: AzĀ isĀ zero, so the āˆ‚Az/āˆ‚y and āˆ‚Az/āˆ‚x derivatives above are also zero, so that leaves only four cross-derivatives to be calculated.

Let’s first define the magnetic dipole moment, however. We talked a lot about it, but what is it? We said it’s similar to the electric dipole momentĀ pĀ = qĀ·d, whose magnitude was Ī»ab = Iab/c2Ā in this example. So what have we got here? No charge, but a current, so that’s chargeĀ per second. No distance either, but some area aĀ·b. And then the 1/c2 factor that, somehow, always slips in because we’ve got it in Maxwell’s equation too. So, we said the fields look the same, and just like Φ was proportional to p, we now see that A is proportional to IĀ·aĀ·b, so it’s quite similar really and so we’ll just define the magnetic dipole moment as:

μ = (current)·(area of the loop) = I·a·b

Now, the area of the loop doesn’t have to be some square: it could also be some circle or some triangle or whatever. You should just change the formula to calculate the area accordingly. You’ll say: where’s theĀ 1/c2 factor? It’s true: we didn’t put it in. Why? I could say: that’s the convention, but that’s a bit too easy as an answer, I guess. I am not quite sure but the best answer I have is that theĀ 1/c2 factor has nothing to do with the physicalityĀ of the situation: it’s part of Maxwell’s equations, or the electromagnetic force equations, if you want. So it’s typical of the magnetic force but it’s got nothing to do with our little loop of current, so to speak. So that’s why we just leave it out in our definition of μ.

OK. That should be it. However, I should also note thatĀ we can (and, hence, weĀ will) make a vector quantity out of μ, so then we write it in boldface: μ. TheĀ symmetry of the situation implies that μ should be some vector that’s normalĀ to the plane of the loop because,Ā if not, all directions would be equally likely candidates. But up or down? As you may expect, we’ll use the right-hand rule again, as illustrated below: if your fingers point in the direction of current, your thumb will point in μ‘s direction. The illustration below also gives the direction of A, because we can now combine the definition of μ and our formulas forĀ Ax, AyĀ and AzĀ and write:

formula magnetic potentialCaptureNote that the formula for A above is quite similar to the formula for Φ below, but not quite the same: in the formula for A we’ve got a vectorĀ crossĀ product – so that’s very different from the vectorĀ dotĀ product below – and then we’ve also got thatĀ 1/c2Ā factor. So watch out: it’s the “same-same but different”, as they say in Asia. šŸ™‚potential formula[…] OK. Sooooo much talk, and still I did notĀ calculateĀ BĀ to effectively show that the magnetic field of a magnetic dipole looks like the electric field of an electric dipole indeed. :-/ So let me no longer postpone those tedious calculations and effectively do thatĀ B =Ā āˆ‡Ć—A cross-productĀ using our new formula for A. Well… Calculating BxĀ and ByĀ is relatively simple because A‘s z-component is zero, so āˆ‚Az/āˆ‚x = āˆ‚Az/āˆ‚y = 0. So for Bx, we get:

Bx

For ByĀ and Bz, we get:

By and z

The … in the formulas above obviously stands for μ/4πε0c2. Now, let’s compare this with the components of the electric field (E) of an electricĀ dipole. Let’s re-write the formula for the electricĀ potential above as:

formula electric dipole

Note that this formula assumes the z-axis points in the same direction as the dipole, and that the dipole is centered at the origin of the coordinate system, so it’s the same coordinate system as the one for the magnetic dipole above. Therefore, we have that cosĪø = z/r, and so the formula above is equivalent to:

formula electric dipole 2

Now, let’s find the electric field by using theĀ EĀ = (Ex, Ey, Ez) = āˆ’āˆ‡Ī¦ = (āˆ’āˆ‚Ī¦/āˆ‚x, Φ, āˆ’āˆ‚Ī¦/āˆ‚y,Ā āˆ’āˆ‚Ī¦/āˆ‚z) equation. ForĀ Ez, we get:

Ez

which we can re-write as:

Ez 2

Huh?Ā Ā Yes. Note that z also appears in r, because rĀ =Ā (x2+y2+z2)1/2, and so we need to apply the product (or quotient) rule for derivatives: (uĀ·v)’ = u’Ā·v + uĀ·v’. Note that this simplifies to EĀ = (1/4πε0)Ā·(2p/r3) for the field at distance r from the dipole in the direction along its axis.

OK, so we’re done withĀ Ez. For ExĀ and Ey, we get:

Ex and EySo… Well… Yes! We’ve checked it: the formulas forĀ Ex, Ey, and EzĀ have exactly the same shape as those for Bx, By, and Bz , except for that 1/c2Ā factor and, of course, the fact that we switched p for μ, so we may have some other number there too. [Oh – before I forget – I promised to give you the formula for the transverse component. The transverse component is, obviously, equal to (Ex2Ā + Ey2)1/2. So it’s just Pythagoras’ rule once more. Let meĀ refer you to Feynman for the explicitĀ formula, i.e. the formula in terms of r andĀ Īø, or in terms of x and y, as I don’t need it here and there’s too much clutter already in this post.]

So… Well… Yes, that’s all very interesting! Indeed, as FeynmanĀ notes, it is quite remarkable that, starting with completely different laws, āˆ‡ā€¢E=ρ/ϵ0Ā andĀ āˆ‡Ć—BĀ =Ā j/ϵ0c2, we end up with the same kind of field.

Is it? Well… Yes and no. You should, of course, note thatĀ the sources whose configuration we summarizing here by the dipole moments are physically quite different—in one case, it’s a circulating current; in the other, a pair of charges—one above and one below the plane of the loop for the corresponding field.Ā So… Well… It’s the same-same but differentĀ indeed! šŸ™‚

Now, to conclude this post, we should, perhaps, have a look at the units and order of magnitude of what we’re talking about here. The unit of charge and the unit of current are related: 1 ampere is a current of 1Ā coulombĀ per second. The coulomb is a rather large unit of charge, as it’s equivalent to the charge of some 6.241Ɨ1018Ā protons. Now, that’s less than a mole, which is aboutĀ 6Ɨ1023, but it’s still a quite respectable number. šŸ™‚ Having said that, a current of one ampere is not so exceptional in everyday life, and so a magnetic dipole moment ofĀ 1 AĀ·m2Ā (i.e. the equivalent of anĀ electric dipole moment of 1 CĀ·m) would not be exceptional either.

The really interesting thing about the magnetic force is that 1/c2Ā factor, which causes its magnitude to be so much smaller than that of the electric force. Indeed, the electricĀ field at distance r from the dipole in the direction along its axis is given by E =Ā (1/4πε0)Ā·(2p/r3). Assuming p = 1 CĀ·m, and noting that 1/4πε0Ā ā‰ˆ 9Ɨ109Ā NĀ·m2/C2 we getĀ E ā‰ˆ (18Ɨ109)Ā·rāˆ’3Ā N/C. So we’ve got an incredibly large (18Ɨ109) factor in front here! [Note how the units come out alright: the electric field is effectively measured inĀ newton per coulomb, indeed.]

What do we get for the magnetic field (along the z-axis once more) from a magnetic dipole with moment μ =Ā 1 AĀ·m2? The formula is now BĀ =Ā (1/4πε0c2)Ā·(2μ/r3). Now, 1/4πε0c2Ā ā‰ˆ (1Ɨ10āˆ’7)Ā NĀ·s2/C2, so B ā‰ˆ (2Ɨ10āˆ’7)Ā·rāˆ’3Ā NĀ·s/CĀ·m.Ā So we’ve got an incredibly smallĀ (2Ɨ10āˆ’7)Ā factor in front here!Ā [As for the unit, 1 NĀ·s/CĀ·m is the unit of the magnetic field indeed. It’s referred to as the tesla (1 T =Ā 1 NĀ·s/CĀ·m), and it makes the force equation come out alright: the magnetic force on some charge q is, indeed, FmagneticĀ = qvƗB, with v expressed in m/s, and so that’s why we need to multiply the N/C factor with an s/m factor in the unit for B.]

So look at this! The point to note is theĀ relative strength – I should sayĀ weakness – of the magnetic force as compared to the electric force: they differ with a factor equal toĀ c2Ā ā‰ˆ 9Ɨ1016. That’s quite incredible, especially because our electric motors and generators work through the magnetic force, as I’ll show in my next posts.Ā In fact, looking at theĀ FmagneticĀ = qvƗB, and comparing it to the electric force (FelectricĀ = qĀ·E), we may say that the magnitude of the magnetic force, as compared to the magnitude of the electric force, isĀ weakerĀ because of two factors:

  1. First, there is the relative velocity of the charge that we are looking, which is equal to β = v/c. The electrostatic Coulomb force has no such factor.
  2. Second, there is the 1/c factor, which we also encountered when discussingĀ radiation pressure, or the momentum of light, which is equal to p = E/c, with the symbol p denoting momentum here – notĀ some dipole moment – and with E denoting the energyĀ of the light photons – not theĀ electricĀ field involved!

Both combine to give us the actual relative strength of the magnetic versus the electric force, so we can write that relative strength as:

|Fmagnetic|/|Felectric| =Ā FB/FEĀ =Ā (v/c)Ā·(1/c) =Ā v/c2

Now, I am tempted to now write a lot about that 1/c factor, but I can’t do that here, as this post has become way too long already. Let me just add a smallĀ addendumĀ here. It’s, perhaps, a not-so-easy piece šŸ™‚ but I warmly recommend trying to work through it.

Addendum: On the momentum of light and radiation pressure

Look at the illustration below: it represents a beam of light, i.e. electromagnetic radiation, originating at some source S and hitting some charge q. What charge? Whatever charge the electromagnetic radiation is affecting, so you may want to think of an electron or a proton in a mirror or a piece of glass, or whatever other surface that is absorbing a photonĀ fully or partially—so that’s any charge that is absorbing the energyĀ of the light.

momentum of light

I’ve written a post on this before. It concluded a series of posts on electromagnetic radiation by noting that we usually look at the electric field only when discussing radiation. The electric field is given by the following formula for E:

formula for E

I know. It looks like a monster, but the first term is just the Coulomb effect. For the explanation of the second and term term, I need to refer you to the mentioned section in Feynman’sĀ Lecture on it. [Sorry for that, but I am still struggling somewhat with this equation myself. If and when I fully ‘get it’ myself, I’ll jot down a summary of my understanding of it. Now now, however.]

The point is: when discussing light, like interference and what have you, we usually forget about the B vector, for which the formula is the following:

magnetic field

So we defineĀ B by referring to E. Of course, this again shows the electric and magnetic force are one and the same phenomenon, really. The cross-product gives you both (i) the magnitude of B, which is equal to B = E/c (the magnitude of the unit vectorsĀ er’ and n in the vector cross-product are obviously equal to one, and sinĪø is equal to one too), and (ii)Ā its direction: it’s perpendicular to both E as well as to er’, which isĀ the unit vector from the pointĀ P where E is measured to the charge q that is producing the field.

The B = E/c equation tells us why the magnetic field is never looked at when discussing electromagnetic radiation: because it’s so tiny. It’s so tiny because of that 1/c factor, and so the B vector in the illustration above is surely notĀ to scale: if all was drawn to scale, you just wouldn’t see it, because it’s 1/c times smaller than E.Ā However, it’s there, and the illustration shows how the magneticĀ force resulting from it looks like, if we forget about scale, that is. šŸ™‚

The magneticĀ force is, obviously,Ā F = qvƗB, and you need to apply the usual right-hand screw rule to find the direction of the force. [Please note that the order of magnitude of the force is the same as that of B so… Well… Again I need to warn you the illustration is not to scale and, hence, somewhat misleading. But I can’t think of an alternative to it.] As you can see, the magnetic force – as tiny as it is – is oriented in the direction of propagation,Ā and it is what is responsible for the so-called radiation pressure.

Indeed, there is a ā€˜pushing momentum’ here, and we can calculate how strong it is. In fact, we shouldĀ check how strong it is in orderĀ to see if it could, potentially, make a space ship like the one below. šŸ™‚

space ship

Sorry. That was just to add a lighter note. šŸ™‚ I know posts like these are rather tedious, so I just wanted to lighten it all up.Ā šŸ™‚ So… Well… Back to ourĀ F = qvƗBĀ equation.Ā Because of that B = –er’×E/c formula, we can substitute B for E/c, and so we get the following formula for the magnitudeĀ of the magnetic force:

F = qĀ·vĀ·E/c

Now, the charge q times the electric field is the electric force on the charge, and the force on the charge times the velocity is equal to dW/dt, so that’s the time rate of change of the work that’s being done on the charge.

Huh?Ā I know: not easy to see what’s being done here. Think of it like this: work equals force times distance, so W = qEĀ·s in this case. Hence, dW/dt = d(qEĀ·s)/dt = qEĀ·ds/dt = qEĀ·ds/dt. [Note that, if you’d check Feynman’s analysis, you’ll find that it’s based on average values here for some reason, so he writes 〈F〉 and 〈E〉, but I am not quite sure why, so I simplified here. I’ll try to figure the why of these averages later, and let you know.]

Again, think physical, rather than mathematical: what’s the charge here? It’s whatever charge the electromagnetic radiation is hitting, so you should think of an electron or a proton in a mirror or a piece of glass or whatever other surface. Now, because of the energy conservation principle,Ā dW/dt must also be equal to theĀ energy that is being absorbed from the light per second.Ā So F = (dW/dt)/c.

Now, Newton’s Law (he had many, of course, so let me be precise: I am talking Newton’sĀ Second Law of Motion)Ā tells us that the force is the rate of change of the momentum: F = mĀ·a = mĀ·dv/dt = d(mĀ·v)/dt. Hence, integrating both sides of theĀ F = (dW/dt)/c equation gives us the associated momentum of the light: p = W/c.

In short, we know that light carries energy, but so we also know that light also carries momentum, and that momentum is due to the magnetic force, and it’s equal to 1/cĀ times the energy. To be clear: it’s real pressure, so when light is emitted from a source there is a recoil effect, and when light hits some charge, it’s the same thing: the momentum of the light is being conserved as the charge that’s absorbing the energy picks it up. To be fully complete: we also have Newton’s Third Law of Motion coming into play: for every action, there is an equal and opposite reaction.

So… Well… That’s about it, I think. Just note that you’ll usually see this momentum written as p = E/c, with E denoting theĀ energyĀ of the radiation, notĀ the electric fieldĀ (sorry for the switch in symbols). In this equation, E isĀ given by the Planck-Einstein relation E = hĀ·f, with h Planck’s constant andĀ fĀ the frequency of the light.Ā 

Interesting! So it all makes sense! Isn’t it wonderful how all these equations come together in the end? šŸ™‚Ā However, I shouldn’t digress even more, and so I’ll leave it to you to further reflect on this. šŸ™‚

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 20, 2020 as a result of a DMCA takedown notice from Michael A. Gottlieb, Rudolf Pfeiffer, and The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/

Magnetostatics: the vector potential

This and the next posts are supposed to wrap up a few loose ends on magnetism. One of these loose ends is the (magnetic) vector potential, which we introduced in our post on gauge transformations, but then we didn’t do much with it. Another topic I neglected so far is that of the magnetic dipole moment (as opposed to the electric dipole moment), which is an extremely important concept both in classical as well as in quantum mechanics. So let’s do the vector potential here, and the magnetic dipole moment in the next. šŸ™‚

Let’s go for it. Let me recall the basics which, as usual, are just Maxwell’s equations.Ā You’ll remember that the electrostatic field was curl-free: āˆ‡Ć—EĀ = 0, everywhere. Therefore, we can apply the following mathematical theorem: if the curl of a vector field is zero (everywhere), then the vector field can be represented as the gradient of some scalar function:

if āˆ‡Ć—CĀ = 0, then there is some ĪØĀ for which C =Ā āˆ‡ĪØ

Substituting C forĀ E, and taking into account our conventions on charge and the direction of flow, we wrote:

E = ā€“āˆ‡Ī¦

Φ (phi) is referred to as the electric potential. Combining E = ā€“āˆ‡Ī¦ with Gauss’ Law – āˆ‡ā€¢E = ρ/ε0Ā āˆ’ we got Poisson’s equation:

āˆ‡2Φ =Ā āˆ’Ļ/ε0

So that equation sums up all of electrostatics. Really: that’s it! šŸ™‚

Now, the two equations for magnetostatics are:Ā āˆ‡ā€¢B = 0 and c2āˆ‡Ć—B = j/ε0. Let me say something more about them:

  1. The āˆ‡ā€¢BĀ = 0 equation is true, always, unlike theĀ āˆ‡Ć—EĀ = 0 expression, which is true for electrostatics only (no moving charges).
  2. The āˆ‡ā€¢BĀ = 0 equation says the divergenceĀ of B is zero,Ā always. Now, you can verify for yourself that the divergence of the curl of a vector field is always zero, so divĀ (curlĀ A) = āˆ‡ā€¢(āˆ‡Ć—A) = 0, always. Therefore, there’s another theorem that we can apply. It says the following: if the divergence of a vector field, say D, is zero – so if āˆ‡ā€¢D = 0, thenĀ D will be theĀ the curl of some other vector fieldĀ C, so we can write:Ā D =Ā āˆ‡Ć—C. Ā Applying this to āˆ‡ā€¢BĀ = 0, we can write:Ā 

If āˆ‡ā€¢BĀ = 0, then there is anĀ A such that B =Ā āˆ‡Ć—A

We can also write this as follows: āˆ‡Ā·B = āˆ‡Ā·(āˆ‡Ć—A) = 0 and, hence, B =Ā āˆ‡Ć—A.Ā Now, it’s this vector field AĀ that is referred to as the (magnetic)Ā vectorĀ potential, and so that’s what we want to talk about here. As a start, it may be good to write all of the components of ourĀ B =Ā āˆ‡Ć—A vector:

formula for B

Note that we have no ‘time component’ because we assume the fields are static, so they do notĀ change with time. Now, because that’s a relatively simple situation, you may wonder whether we really simplified anything with this vector potential. B is a vector with three components, and so is A. The answer to that question is somewhat subtle, and similar to what we did for electrostatics: it’sĀ mathematically convenient to use A, and then calculate the derivatives above to find B. So the number of components doesn’t matter really: it’s just more convenientĀ to first get A using our data on the currents j, and then we get B from A.

That’s it really. Let me show you how it works. The whole argument is somewhat lengthy, but it’s not difficult, and once it’s done, it’s done. So just carry on and please bear with me šŸ™‚

First, we need to put some constraints on A, because the B =Ā āˆ‡Ć—A equation does not fully define A. It’s like the scalar potential Φ: any Φ’ = Φ + C was as good a choice as Φ (withĀ CĀ any constant), so we needed a reference point Φ = 0, which we usually took at infinity. With the vector potentialĀ A, we have even more latitude: we can not only add a constant but any field which is the gradient of some scalar field, so any A’ = A + āˆ‡ĪØ will do. Why? Just write it all out: āˆ‡Ć—(A + āˆ‡ĪØ) = āˆ‡Ć—A + āˆ‡Ć—(āˆ‡ĪØ). But the curl of the gradient of a scalar field (or a scalar function) is always zero (you can check my post on vector calculus on this), so āˆ‡Ć—(āˆ‡ĪØ) = 0 and soĀ āˆ‡Ć—(A + āˆ‡ĪØ) = āˆ‡Ć—A + āˆ‡Ć—(āˆ‡ĪØ) = āˆ‡Ć—A + 0 = āˆ‡Ć—A = B.

So what constraints should we put on our choice of A? The choice is, once again, based onĀ mathematical convenience: in magnetostatics,Ā we’ll choose A such that āˆ‡ā€¢AĀ = 0. Can we do that? Yes. The A’ = A + āˆ‡ĪØ flexibility allows us to make āˆ‡ā€¢A’Ā anything we wish, and so A and A’ will have the same curl, but theyĀ don’t need to have the same divergence. So we can choose an A’ soĀ āˆ‡ā€¢A’Ā = 0, and then we denote A’ by A. šŸ™‚ So our ‘definition’ of the vector potentialĀ A is now:

B =Ā āˆ‡Ć—A andĀ āˆ‡ā€¢AĀ = 0

I have to make two points here:

  1. First, you should note that, in my post on gauges, I mentioned that the choice is different when the time derivatives of E and B are notĀ equal to zero, so when we’re talking changingĀ currents and charge distributions, so that’sĀ dynamics. However,Ā that’s not a concern here.
  2. To be fully complete, I should note that the ‘definition’ above does still notĀ uniquelyĀ determine A. For a unique specification, we also need some reference point, or say how the field behaves on some boundary, or at large distances. It is usually convenient to choose a field which goes to zero at large distances, just like our electricĀ potential.

Phew! We’ve said so many things about AĀ now, but nothingĀ that has any relevance to how we’d calculateĀ A. 😦 So we are we heading here?

Fortunately, we can go a bit faster now. TheĀ c2āˆ‡Ć—B = j/ε0Ā equation and our B =Ā āˆ‡Ć—A give us:

c2āˆ‡Ć—(āˆ‡Ć—A)Ā = j/ε0

Now, there’s this other vector identity, which you surely won’t remember either—but trust me: I am not lying: āˆ‡Ć—(āˆ‡Ć—A) = āˆ‡(āˆ‡ā€¢A) āˆ’Ā āˆ‡2A. So, now you see why we choose A such that āˆ‡ā€¢A = 0 ! It allows us to write:

c2āˆ‡Ć—(āˆ‡Ć—A)Ā = āˆ’ c2āˆ‡2A =Ā j/ε0Ā ā‡”Ā āˆ‡2A = –j/ε0c2

Now, the three components ofĀ āˆ‡2A = –j/ε0c2Ā are, of course:

components

As you can see, each of these three equations is mathematically identical to that Poisson equation:Ā āˆ‡2Φ =Ā āˆ’Ā Ļ/ε0. So all that we learned about solving for potentials when ρ is known can now be used to solve for each component of A when j is known. Now, to calculate Φ, we used the following integral:

formula 1Simply substituting symbols then gives us the solution forĀ Ax:

formula 2

We have a similar integral for AyĀ and Az, of course, and we can combine the three equations in vector form:

formula 3

Finally, and just in case you wonder what is what, there’s the illustration below (taken from Feynman’s Lecture on this topic here) that, hopefully, will help you to make sense of it all.

illustration A

At this point, you’re probably tired of these formulas (or asleep) or (if you’re not asleep) wondering what they mean really, so let’s do two examples. Of course, you won’t be surprised that we’ll be talking a straight wire and a solenoid respectively once again. šŸ™‚

The magnetic field of a straight wire

We already calculated the magnetic field of a straight wire, using AmpĆØre’s Law and the symmetry of the situation, in our previous post on magnetostatics. We got the following formula:

B formula

Do we get the same using those formulas for A and then doing our derivations to get B? We should, and we do, but I’ll be lazy here and just refer you to the relevant section in Feynman’s LectureĀ on it, because the solenoid stuff is much more interesting. šŸ™‚

The magnetic field of a solenoid

In the mentioned post on magnetostatics, we also derived a formula for the magnetic field insideĀ a solenoid. We got:

solenoid formula 2with nĀ the number of turns per unit length of the solenoid, and I the current going through it. However, in the mentioned post, we assumed that the magnetic field outside of the solenoid was zero, for all practical purposes, but it is not. It is very weak but not zero,Ā as shown below. In fact, it’s fairly strong at very short distances from the solenoid! Calculating the vector potential allows us to calculate its exact value, everywhere. So let’s go for it.

solenoid 2

The relevant quantities are shown in the illustration below. So we’ve got a very long solenoid here once again, with n turns of wire per unit length and, therefore, a circumferential current on the surface of nĀ·I per unit lengthĀ (the slight pitch of the winding is being neglected).

solenoid

Now, just like that surface charge density ρ in electrostatics, we have a ‘surface current density’ J here, which we define as J =Ā nĀ·I. So we’re going from a scalar to a vector quantity, and the components of J are:

JxĀ = –JĀ·sinĻ•, JyĀ = –JĀ·cosĻ•, JzĀ = 0

So how do we do this? As should be clear from the whole development above, the principle is that the x-component of the vector potential arising from a current density j is the same as the electric potential Φ that would be produced by a charge density ρ equal to jxĀ divided by c2, and similarly for the y- and z-components.Ā Huh?Ā Yes. Just read it a couple of times and think about it: we should imagine some cylinder with a surface charge ρ = –(J/c2)Ā·sinĻ• to calculate Ax. And then we equate ρ with –(J/c2)Ā·cosĻ• and zero respectively to find AyĀ and Az. Ā 

Now, that sounds pretty easy but Feynman’s argument is quite convoluted here, so I’ll just skip it (click the link here if you’d want to see it) and give you the final result, i.e. the magnitude of A:

vector potential

Of course, you need to interpret the result above with the illustration, which shows that A is always perpendicular to r’. [In case you wonder why we write r’Ā (so r with a prime) and not r, that’s to make clear we’re talking the distance from the z-axis, so it’sĀ not the distanceĀ from the origin.]

Now, you may think that c2Ā in the denominator explains the very weak field, but it doesn’t: it’s the inverse proportionality to r’ that makes the difference!Ā Indeed, you should compare the formula above with the result we get for the vector potential insideĀ of the solenoid, which is equal to:

A inside

The illustration below shows the quantities involved. Note that we’re talking aĀ uniformĀ magnetic field here, along the z-axis, which has the same direction as B0Ā and, hence, is pointing towards you as you look at the illustration, which is why you don’t see the B0Ā field lines and/or the z-axis: they’re perpendicular to your computer screen, so to speak.

illustration A insideAs for the direction of A, it’s shown on the illustration, of course, but let me remind you of the right-hand rule for the vector cross product aƗb once again, so you can make sense of the direction of AĀ = (1/2)B0Ɨr’ indeed:

right-hand ruleAlso note the magnitude this formula implies:Ā aƗb = |a|Ā·|b|Ā·sinĪøĀ·n, withĀ Īø the angle between a and b,Ā andĀ nĀ the normal unit vector in the direction given by that right-hand rule above. Now, unlike a vectorĀ dotĀ product, the magnitude of the vector cross product isĀ notĀ zero for perpendicular vectors. In fact, whenĀ Īø = π/2, which is the case for B0Ā andĀ r’, then sinĪø = 1, and, hence, we can write:

|A| = A =Ā (1/2)|B0||r’| = (1/2)Ā·B0Ā·r’

Now, just substitute B0 for B0 = n·I/ε0c2, which is the field inside the solenoid, then you get:

A = (1/2)Ā·nĀ·IĀ·r’/ε0c2

You should compare this formula with the formula for A outsideĀ the solenoid, so you can draw the right conclusions. Note that both formulas incorporate the same (1/2)Ā·nĀ·I/ε0c2Ā factor. The difference, really, is that insideĀ the solenoid, A is proportional to r’ (as shown in the illustration: if r’ doubles, triples etcetera, then A will double, triple etcetera too) while, outside of the solenoid, A is inverselyĀ proportional to r’. In addition, outsideĀ the solenoid, we have the a2Ā factor, which doesn’t matter inside. Indeed, the radius of the solenoid (i.e. a) changes the flux, which is the product of B and the cross-section area π·a2, but not B itself.

Let’s do a quick check to see if the formula makes sense. We do not want A to be larger outside of the solenoid than inside, obviously, so the a2/r’ factor should be smallerĀ than r’ for r’ > a. Now,Ā a2/r’ <Ā r’ if a2 < r’2, and because a an r’ are both positive real numbers, that’s the case if r’ >Ā a indeed. So we’ve got something that resembles the electric field inside and outside of a uniformly charged sphere, except that A decreases as 1/r’ rather than as 1/r’2, as shown below.

graph
Hmm… That’s all stuff to think about…Ā The thing you should take home from all of this is the following:

  1. A (uniform) magnetic field B in the z-direction corresponds to a vector potential A that rotates about the z-axis with magnitudeĀ A =Ā B0Ā·r’/2 (with r’ the displacement from the z-axis,Ā notĀ from the origin—obviously!). So that gives you the A insideĀ of a solenoid. The magnitude isĀ A = (1/2)Ā·nĀ·IĀ·r’/ε0c2, so A is proportional with r’.
  2. Outside of the solenoid, A‘s magnitude (i.e. A) is inverselyĀ proportional to the distance r’, and it’s given by the formula: A = (1/2)Ā·nĀ·IĀ·a2/ε0c2Ā·r’. That’s, of course, consistent with the magnetic field diminishing with distance there. But remember: contrary to what you’ve been taught or what you often read, it isĀ notĀ zero. It’s onlyĀ nearĀ zero if r’ >> a.

Alright. Done. Next post. So that’s on the magnetic dipole moment šŸ™‚

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/

Ferroelectrics and ferromagnetics

Ferroelectricity and ferromagnetism are two different things, but they are analogous. Materials are ferroelectric if they have a spontaneous electric polarization that can be changed or reversed by the application of an external electricĀ field.Ā Ferromagnetism, in contrast, refers to materials which exhibit a permanentĀ magneticĀ moment.

The materials are very different.Ā In fact, most ferroelectric materials do not contain any iron and, hence, the ferroĀ in the term is somewhat misleading. Ferroelectric materials are a special class ofĀ crystals, like barium or lead titanate (BaTiO3Ā or PbTiO3). Lead zirconate titanate (LZT) is another example. These materials are alsoĀ piezoelectric: when applying some mechanical stress, they will generate some voltage. In fact, the process goes both ways: when applying some voltage to them, it will also create mechanical deformation, as illustrated below (credit for this illustration goes to Wikipedia).

Piezo_bending_principle

Ferroelectricity has to do withĀ electricĀ dipoles, while ferromagnetism has to do withĀ magneticĀ dipoles. We’ve only discussed electric dipolesĀ so far (see the section on dielectrics in my post on capacitors) and so we’re only in a position to discuss ferroelectricity right now, which is what I’ll do here. However, before doing so, let me briefly quote from the Wikipedia article on ferromagnetism, because that’s really concise and to the point on this:

“One of the fundamental properties of an electron (besides that it carries charge) is that it has a magnetic dipole moment, i.e. it behaves like a tiny magnet. This dipole moment comes from the more fundamental property of the electron that it has quantum mechanical spin. Due to its quantum nature, the spin of the electron can be in one of only two states; with the magnetic field either pointing “up” or “down” (for any choice of up and down). The spin of the electrons in atoms is the main source of ferromagnetism, although there is also a contribution from the orbital angular momentum of the electron about the nucleus.”

In short, ferromagnetismĀ was discovered and known much before ferroelectricity was discovered and studied, but it’s actually more complicated, because it’s a quantum-mechanical thing really, unlike ferroelectricity, which we’ll discuss now. Before we start, let me note that, in many ways, this post is a continuation of the presentation on dielectrics, which I referred to above already, so you may want to check that discussion in that post I referred toĀ if you have trouble following the arguments below.

Molecular dipoles

Let me first remind you of the basics. The (electric)Ā dipole moment is the product of the distance between two equal but opposite charges q+Ā and qāˆ’. Usually, it’s written as a vector so as to also keep track of its directionĀ and use it inĀ vector equations, so we write pĀ = qd, with d the vector going from the negative to the positive charge, as shown below.

dipole_limit256

Now, molecules like water molecules have a permanent dipole moment, as illustrated below. It’s because the center of ‘gravity’ of the positive and negative charges doĀ notĀ coincide, so that’s what makes the H2O molecule polar, as opposed to the O2Ā molecule, which is non-polar.

molecular dipoles

Now, if we placeĀ polarĀ molecules in some electric field, we’d expect them to line up, to some extent at least, as shown below (the second illustration hasĀ moreĀ dipoles pointing vaguely north).

Capture a

Capture b

However, at ordinary temperatures and electric fields, the collisions of the molecules in their thermal motion keeps them from lining up too much. In fact, we can apply the principles of statistics mechanics to calculate how much exactly. You can check out the details in Feynman’sĀ LectureĀ on it, but the result is that theĀ net dipole moment per unit volumeĀ (so that’s the polarization)Ā is equal to:

polarization

So the polarization is proportional to the number of molecules per unit volume (N), the square of their dipole moment (p0) and, as we’d might expect, the electric field E, and inversely proportional to the temperature (T). In fact, the formula above is a sort of first-order approximation, in line with what we wrote on the electric susceptibility χ (chi) in our post on capacitors, where we also assumed the relation between P and E was linear, so we wrote:Ā P = ε0·χ·E. Now, engineers and physicists often use different symbols and definitions and so you may of may not have heard about another concept saying essentially the same thing: the dielectric constant, which is denoted byĀ ĪŗĀ (kappa) and is, quite simply, equal to Īŗ =Ā 1 + χ. Combining the expression for P above, and theĀ P = ε0·χ·E = ε0Ā·(Īŗāˆ’1)Ā·E expression, we get:

susceptibility

This doesn’t say anything new: it just states the dependence of χ on the temperature. Now, you can imagine this linear relationship has been verified experimentally. As it turns out, it’s sort of valid, but it is notĀ as straightforward as you might imagine. There’s a nice post on this on the University of Cambridge’s Materials Science site. But this blog is about physics, not about materials science, so let’s move on. The only thing I should add to this section is a remark on the energy of dipoles.

You know charges in a field have energy,Ā potentialĀ energy. You can look up the detail behind the formulas in one of my other posts on electromagnetism, so I’ll just remind you of them: the energy of a charge is, quite simply, the product of the charge (q) and the electricĀ potentialĀ (Φ) at the location of the charge. Why? Well… The potential is the amount of work we’d do when bringing the unitĀ charge there from some other (reference) point where Φ = 0. In short, the energy of the positive charge is q·Φ(1) and the energy of the negative charge isĀ āˆ’q·Φ(2), with 1 and 2 denoting their respective location, as illustrated below.

energy dipole

So we have U =Ā q·Φ(1)Ā āˆ’Ā q·Φ(2) = qĀ·[Φ(1)āˆ’Ī¦(2)]. Now, we’re talking tiny little dipoles here, so we can approximate ΔΦ = Φ(1)āˆ’Ī¦(2) by ΔΦ = dā€¢āˆ‡Ī¦ = Ī”xĀ·(āˆ‚Ī¦/āˆ‚x) + Ī”yĀ·(āˆ‚Ī¦/āˆ‚y). Hence, also noting that E = āˆ’āˆ‡Ī¦ and qd = p0, we get:

U =Ā q·Φ(1)Ā āˆ’Ā q·Φ(2) = qdā€¢āˆ‡Ī¦ = Ā āˆ’p0•E = āˆ’p0Ā·EĀ·cosĪø, withĀ Īø the angle between p0Ā andĀ E

So the energy is lowerĀ when the dipoles are lined up with the field, which is what we would expect, of course. However, it’s an interesting thing so I just wanted to show you that. šŸ™‚

Electrets, piezoelectricity and ferroelectricity

The analysis above was very general, so we actually haven’t started our discussion on ferroelectricity yet! All of the above is just a necessary introduction to the topic. So let’s move on. Ferroelectrics are solids, so let’s look at solids. Let me just copy Feynman’s introduction here, as it’s perfectly phrased:

“The first interesting fact about solids is that there can be a permanent polarization built in—which exists even without applying an electric field. An example occurs with a material like wax, which contains long molecules having a permanent dipole moment. If you melt some wax and put a strong electric field on it when it is a liquid, so that the dipole moments get partly lined up, they will stay that way when the liquid freezes. The solid material will have a permanent polarization which remains when the field is removed. Such a solid is called an electret.Ā An electret has permanent polarization charges on its surface. It is the electrical analog of a magnet. It is not as useful, though, because free charges from the air are attracted to its surfaces, eventually cancelling the polarization charges. The electret is ā€œdischargedā€ and there are no visible external fields.”

Another example (i.e. other than wax) of an electret is the crystal lattice below. As you can see, all the dipoles are pointing in the same direction even with no applied electric field. Many crystals have such polarization but, again, we do not normally notice it because the external fields are discharged, just as for the electrets.

crystalNow, this gives rise to the phenomena ofĀ pyroelectricityĀ andĀ piezoelectricity. Indeed, as Feynman explains: “If these internal dipole moments of a crystal are changed, external fields appear because there is not time for stray charges to gather and cancel the polarization charges. If the dielectric is in a condenser, free charges will be induced on the electrodes. The moments can also change when a dielectric is heated, because of thermal expansion. The effect is called pyroelectricity. Similarly, if we change the stresses in a crystal—for instance, if we bend it—again the moment may change a little bit, and a small electrical effect, called piezoelectricity, can be detected.”

But, still, piezoelectricity is not the same as ferroelectricity. In fact, there’s a hierarchy here:

  1. Out of all crystals, some will be piezoelectric.
  2. Among all piezoelectric crystals, some will also be pyroelectric.
  3. Among the pyroelectric crystals, we can find some ferroelectricĀ crystals.

The defining characteristic of ferroelectricity is that the built-in permanent moment can be reversed by the application of an external electric field. Feynman defines them as “nearly cubic crystals, whose moments can be turned in different directions, so we can detect a large change in the moment when an applied electric field is changed: all the moments flip over and we get a large effect.”

Because this is a blog,Ā notĀ a physics handbook, I’ll refer you to FeynmanĀ and/or the Wikipedia article on ferroelectricity for an explanation of the mechanism. Indeed, the objective of this post is to explainĀ whatĀ it is, and so I don’t want to go off into the weeds. The two diagrams below, which I took from the mentioned Wikipedia article, illustrate the difference between your average dielectric material as opposed to a ferroelectric material. The first diagram shows you the linear relationship between P and E we discussed above: if we reverse the field, so E becomes negative, then the polarization will be reversed as well, but gradually, as shown below.

160px-Dielectric_polarisation

In contrast, the illustration below shows a hysteresis effect,Ā which can be used as a memory function, and ferroelectric materials are indeed used for ferroelectric RAM (FeRAM) memory chips for computers! I’ll let you googleĀ that for yourself āˆ’ it’s fun: just have a look at the following link, for example āˆ’ because it’s about time I start wrapping up this post. šŸ™‚

160px-Ferroelectric_polarisationOK. That’s it for today. More tomorrow. šŸ™‚

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 20, 2020 as a result of a DMCA takedown notice from Michael A. Gottlieb, Rudolf Pfeiffer, and The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 20, 2020 as a result of a DMCA takedown notice from Michael A. Gottlieb, Rudolf Pfeiffer, and The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/

Magnetostatics

Pre-script (dated 26 June 2020): This post got mutilated by the removal of some material by the dark force. You should be able to follow the main story line, however. If anything, the lack of illustrations might actually help you to think things through for yourself.

Original post:

Not all is exciting when studying physics. In fact, electromagnetism is, most of the time, a extremelyĀ boringĀ subject-matter. But so we need to get through it, because we need the math and the formulas. So… Here we go… :-/

When going from electrostatics to electrodynamics, one first needs to have a look at magnetostatics, to get familiar with (steady)Ā electric currents. So let’s have a look at what they are. Of course,Ā you already know what steady currents are. In that case, you should, perhaps, stop reading. But I’d recommend you go through it anyway. It’s always good to be explicit, so let’s be explicit.

Let me first make a very pedantic note. There areĀ a couple of sections in Feynman’s LecturesĀ in which he assumes that a steady current in a wire is uniformly distributed throughout the cross-section of the current-carrying wire: that assumption amounts to saying that the current densityĀ j is uniform. He uses that assumption, for example, when calculating the force per unit length of a current-carrying wire in a magnetic field (see Vol. II, section 13-3). He also uses it when calculating the magnetic field it creates itself (see Vol. II, section 14-3). This raises two questions:

  1. Is the assumption true?
  2. Does it matter?

My impression is that it’s a simplification that doesn’t matter. So the answer to both question would be negative. But let’s examine them. First note that, in previous posts, we repeatedly said that, if we place a charge Q on any conductor, all charges will spread out in some way on the surface, so we have an equipotential on the surface and no electric fieldĀ insideĀ of the conductor. The physics behind are easy to understand: if there were an electric field inside of the conductor, and the surface were not an equipotential, the charges would keep moving until it became zero.

Does it matter? Maybe. Maybe not. I discussed the electric field from a conductor in a previous post, so let me just recall some formulas here, first and foremost Gauss’ Law, which says that the electricĀ flux from any closed surface S is equal to Qinside/ε0. Now,Ā QinsideĀ is, obviously, the sum of the charges inside the volume enclosed by the surface, and the most remarkable thing about Gauss’ Law is that the chargeĀ distribution inside of the volumeĀ doesn’t matter. So if we’re talking a uniformly charged sphere or a thin spherical shellĀ ofĀ charge, it’s the same. The illustration below shows the field for a uniformly charged sphere: E is proportional to rĀ (to be precise: E =Ā (ρ·r)/(3ε0) for r ≤ R) insideĀ the sphere,Ā and outsideĀ E is proportional to 1/r2Ā (to be precise: E = Qinside/(4πε0r2) for r ≄ R).

Field of a sphere

However, Gauss’ Law is a law that gives us theĀ electric fluxĀ only, so we’re talking E only. We also have the magnetic field, i.e. the field vectorĀ B. So what’s the equivalent of Gauss’ Law for B? That’s AmpĆØre’s Law, obviously, so let’s have a look at how Feynman derives that law.

AmpĆØre’s Law

Feynman starts by defining the current through some surface S as the following integral:

Integral for I

The illustration below explains the logic behind. The vector j is like the heat flow vectorĀ hĀ which we used when explaining the basics of vector calculus: it is some amount passing expressedĀ per unit time and per unit area.Ā As for the use of n, that’s the same normal unit vector we used for h as well: we then wrote thatĀ hĀ·nĀ =Ā |h|Ā·|n|Ā·cosĪø = hĀ·cosĪø was theĀ componentĀ of the heat flow that’s perpendicular or normal (as mathematicians prefer to say) to the surface. So here we’ve got the same: jĀ·nĀ·dS is the amount of charge flowing across an infinitesimally small area dS in a unit time. So to get the electric current I, which isĀ the total charge passing per unit time through a surface S, we need to integrate the normal component of the flow through all the surface elements, which is what the integral above is doing.

Illustration INote that I isĀ notĀ a vector but a scalar. We could, however, include the idea of the direction of flow byĀ making I a vector, so then we write it in boldface:Ā I.Ā It is measured inĀ coulombĀ per second, aka asĀ ampere: 1 A = 1 C/s. Also note we don’t have any wires here: just surfaces and volumes. šŸ™‚ Onwards!

The equations ofĀ magnetostaticsĀ are Maxwell’s third and fourth equation and,Ā as we used Maxwell’s first and second equation to derive Gauss’ Law, we’ll use these two to derive AmpĆØre’s Law: (1)Ā āˆ‡ā€¢B = 0 and (2) c2āˆ‡Ć—B = j/ε0.

You know these equations: the first one basically says there’s no flux of B: there’s no such thing asĀ magneticĀ charges, in other words. The second one says that a current produces someĀ circulationĀ of B. You also know these equations are valid only forĀ staticĀ fields: all electric charge densities are constant, and all currents are steady, so the electric and magnetic fields are not changing with time:Ā āˆ‚E/āˆ‚t = 0 =Ā āˆ‚B/āˆ‚t. Forget about c2Ā for a moment (it’s just a constant) and note thatĀ āˆ‡Ć—BĀ is referred to as the curlĀ of B.

Now, as I pointed out in one of my posts on vector analysis, the divergenceĀ of the curlĀ of a vector is always equal to zero, so āˆ‡ā€¢(āˆ‡Ć—B) = 0. However, because āˆ‡Ć—B = j/ε0c2, that means āˆ‡ā€¢(j/ε0c2) must also be equal to zero (we’re just taking the divergence of both sides of the equation here), and so we find that āˆ‡ā€¢jĀ must be equal to zero. What does that mean?Well… From the same post, you may or may not remember that theĀ divergenceĀ of some vector field C (so that’s āˆ‡ā€¢C) is the (net)Ā fluxĀ out of an (infinitesimal) volume around the point we’re considering, soĀ āˆ‡ā€¢jĀ = 0 implies that as much charge must be coming in as it going out,Ā always and everywhere. So that means that, because of the charge conservation law (no charges are created or lost), we can only look at charges flowing in paths that close back on themselves, so we can only consider closed circuits. It’s a minor point – so don’t worry too much about it – but it does imply that we’re not looking at condensers,Ā for example. Just remember: magnetostatics is aboutĀ circulation, we have no flux, not of B, and not of j: our field, and our charges, circulate.Ā šŸ™‚

OK. Let’s get back to the lesson. We need to find AmpĆØre’s Law, so we’d better get on with it. šŸ™‚ To find Gauss’ Law, we used Gauss’ Theorem. To find AmpĆØre’s Law, we’ll use… Stokes’ Theorem. [Sorry!] I need to refer you, once again, to that post on vector analysis for it. Here I can only remind you of the Theorem itself. It says thatĀ the line integral of the tangential component of a vector (field) around a closed loop is equal to the surface integral of the normal component of the curl of that vector over any surface which is bounded by the loop. […] I know that’s quite a mouthful, so let me jot down the equation:

Stokes Theorem-2

Applying it to the magnetic field vector B, we get:

Stokes Theorem for B

This is the illustration which goes with it.

Stokes Theorem for B graph

Now, using ourĀ āˆ‡Ć—B = j/ε0c2Ā equation, we get:

Integral 2

Finally, we just plug in our I = ∫ jĀ·nĀ dS integral and we’re done. This is AmpĆØre’s Law:

Ampere's Law

It basically says that the circulation of B around any closed curve is equal to the current I through the loop, divided by ε0c2. So what can we do with it? Well…Ā We used Gauss’ Law to find the electric field in various circumstances, so let’s now use AmpĆØre’s Law to find theĀ magneticĀ field in various circumstances. šŸ™‚

Before doing so, however, let me note thatĀ AmpĆØre’s Law doesĀ notĀ depend on any particular assumption in regard to the distribution of the charge densities j. So, frankly speaking, don’t worry too much about that assumption about a steady current in a wire: a current is a current inĀ AmpĆØre’s Law. šŸ™‚

Wires

You know the magnetic field around a wire, as you’ll surely remember that right-hand rule for it from your high-school physics classes.Ā Note, however, that it assumes you apply the usual convention: charge flows fromĀ positiveĀ toĀ negative, because our unit of electric charge is obviously +1, not –1. So theĀ electron flowĀ actually goes the other way. šŸ™‚

right-hand rule

But so we’re past our high school days and we need to apply AmpĆØre’s Law. The symmetry of the situation implies that that line integral of BĀ·ds, taken along some closedĀ circleĀ around the wire, is, quite simply, the magnitude of B times the circumference r of our circle. Indeed, the symmetry of the situation implies that BĀ at some distance r should be of the same magnitude everywhere, so we have:

B

But from AmpĆØre’s Law we know that integral is equal to I/ε0c2Ā and, therefore, BĀ·2π·r must equalĀ I/ε0c2, and so we get the grand result we were looking for. The magnetic field outside of a (long) wire carrying the current I is:

B formula

As Feynman notes, we can write this in vector form to include the directions, remembering that B is at right angles both to I as well as to r, and remembering that the order matters, of course, because of the right-hand rule for a vector cross product. šŸ™‚

Wire formula 507px-right_hand_rule_cross_product

Solenoids

Coils of wire, and solenoids, pop up almost everywhere when studying electromagnetism. Indeed, transformers, inductances, electrical motors: it’s all coils. So, yes, we can’t escape them. 😦 So let’s get on with it. As you know, aĀ solenoid is a long coil of wire wound in a tight spiral. The illustrations below show a cross-section and its magnetic field.

solenoid 2 solenoid 1

Now, this is probably one of Feynman’s most intuitive arguments. Read: he’s cutting an awful lot of corners here. šŸ™‚ I’ll just copy him:

We observe experimentally that when a solenoid is very long compared with its diameter, the field outside is very small compared with the field inside. Using just that fact, together with AmpĆØre’s law, we can find the size of the field inside.Ā Since the field stays inside (and has zero divergence), its lines must go along parallel to the axis, as shown above. That being the case, we can use AmpĆØre’s law with the rectangular ‘curve’ Ī“ shown in the figure. This loop goes the distanceĀ LĀ inside the solenoid, where the field is, say, B0, then goes at right angles to the field, and returns along the outside, where the field is negligible. The line integral of BĀ for this curve is just B0Ā·L,Ā and it must beĀ 1/ε0c2Ā times the total current through Ī“, which isĀ NĀ·IĀ if there areĀ NĀ turns of the solenoid in the lengthĀ L. We have:

solenoid formula 1

Or, lettingĀ nĀ be the number of turns per unit length of the solenoid (that is, n=N/L), we get:

solenoid formula 2

Oh… What happens to the lines of B when they get to the end of the solenoid? Well… They just spread out in some way and return to enter the solenoid at the other end. Hmm… He’s really cutting corners here, isn’t he? But the formula is right, and I’d rather keep it short—just like he seems to want to do here. šŸ™‚ I’ll just insert an illustration showing another right-hand rule—the right-hand rule for solenoids: if the direction of the fingers of your right hand is the direction of current, then your thumb gives the direction of the magnetic field inside.

helix1

You may wonder: does it matter where the +Ā and āˆ’ ends of the coil are? Good question because, in practice, we’ll have something that’s very tightly wound, like the coil below, so when making an actual coil (click on this link for a nice video), we’ll have several rows and so we wind from right to left and then back from left to right and so on and so on. So if we’d have two rows of wire, the two ends of the wire would come out on the same side, and that’s OK.

OLYMPUS DIGITAL CAMERA

Of course, the wire needs to be insulated. What you see on the picture (and in the video) is the use of so-calledĀ magnet wire, which has a polymer film insulation. So when making the electrical connections at both ends, after winding the coil, you need to get rid of the insulation, but then it often melts just by the heat of soldering. And now that we’re talking practical stuff, let me say something about the magnetic core you see in the illustration above.

A magnetic coreĀ is a material with high magneticĀ permeabilityĀ as compared to the surrounding air, and this high permeability will cause the magnetic field to be concentrated in the core material. Now, there’s a phenomenon that’s calledĀ hysteresis, which means that the core material will tend to retain itsĀ magnetization when the applied field is removed. This is not very desirable in many applications, such as transformers or electric engines. That’s why so-called ‘soft’ magnetic materials with low hysteresis are often preferred. The so-called softĀ iron is such material: it’s literally softer because of a heat treatment increasing its ductility and reducing its hardness. Of course, for permanent magnets, a so-called ‘hard’ magnetic material will be used. But here we’re getting into engineering and that’s not what I want to write about in this blog.

I’ll just end by noting that aĀ magnetic field has a so-called north (N)Ā andĀ south (S) pole. That convention refers to the Earth’s north and south pole, of course. However, since opposite poles (north and south) attract, the North Magnetic Pole is actually the south pole of the Earth’s magnetic field, and the South is the north. šŸ™‚ So it’s better not to think too much of the Earth’s poles when discussing the poles of a magnet. By convention, aĀ magnet’s north pole is where the field lines of a magnet emerge, and the south pole is where they enter, as shown below.

220px-VFPt_cylindrical_magnet_thumb

In any case… Folks: that’s it for today. I’ll continue tomorrow. šŸ™‚

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 20, 2020 as a result of a DMCA takedown notice from Michael A. Gottlieb, Rudolf Pfeiffer, and The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/

A post for Vincent: on the math of waves

Pre-scriptum (dated 26 June 2020): These posts on elementary math and physics for my kids (they are 21 and 23 now and no longer need such explanations) have not suffered much the attack by the dark force—which is good because I still like them. While my views on the true nature of light, matter and the force or forces that act on them have evolved significantly as part of my explorations of a more realist (classical) explanation of quantum mechanics, I think most (if not all) of the analysis in this post remains valid and fun to read. In fact, I find the simplest stuff is often the best. šŸ™‚

Original post:

I wrote this post to just briefly entertain myself and my teenage kids. To be precise, I am writing this for Vincent, as he started to study more math this year (eight hours a week!), and as he also thinks he might go forĀ engineering studies two years from now. So let’s see if he gets this and āˆ’ much more importantlyĀ āˆ’Ā if he likes the topic. If not… Well… Then he should get even better at golf than he already is, so he can make a living out of it. šŸ™‚

To be sure, nothing what I write below requires an understanding of stuff you haven’t seen yet, like integrals, or complex numbers. There’s no derivatives, exponentials or logarithms either: you just need to know what a sine or a cosine is, and then it’s just a bit of addition and multiplication. So it’s just… Well… Geometry and waves as I would teach it to an interested teenager. So let’s go for it.Ā And, yes,Ā I am talking to you now, Vincent!Ā šŸ™‚

The animation below shows a repeatingĀ pulse. It is a periodic function: a traveling wave. It obviously travels in the positive x-direction, i.e. from left to right as per our convention. As you can see, theĀ amplitudeĀ of our little wave varies as a function of time (t) and space (x), so it’s a function in two variables, like y = F(u, v). You know what that is, and you also know we’d refer to y as the dependent variable and to u and v as the independentĀ variables.

pulse

Now, because it’s a wave, and because it travels in the positive x-direction, the argument of the wave function FĀ will be xāˆ’ct, so we write:

y = F(xāˆ’ct)

Just to make sure: c isĀ the speed of travel of this particular wave, so don’t think it’s the speed of light. This wave can be any wave: a water wave, a sound wave,… Whatever. Our dependentĀ variableĀ y is the amplitude of our wave, so it’s the vertical displacementĀ āˆ’ up or down āˆ’ of whatever we’re looking at. As it’s a repeating pulse, y is zero most of the time, except when that pulse is pulsing. šŸ™‚

So what’s the wavelength of this thing?

[…] Come on, Vincent. Think! Don’t just look at this!Ā 

[…] I got it, daddy! It’s the distance between two peaks, or between the center of two successive pulses— obviously! šŸ™‚

[…]Ā Good! šŸ™‚ OK. That was easy enough.Ā Now look at the argument of this function once again:

F = F(xāˆ’ct)

We are not merely acknowledging here that F is some function of x and t,Ā i.e. some function varying in space and time. Of course, F is that too, so we can write:Ā y = F = F(x, t) = F(xāˆ’ct), but it’sĀ moreĀ than just some function: we’ve got a very special argument here,Ā xāˆ’ct, and so let’s start our little lesson by explaining it.

The xāˆ’ct argument is there because we’re talking waves, so that is something moving through space and time indeed. Now, what are we actuallyĀ doing when we write xāˆ’ct? Believe it or not, we’re basically converting something expressed in time units into something expressed in distance units. So we’re converting time into distance, so to speak. To see how this works, suppose we addĀ some timeĀ Ī”t to the argument of our function y = F, so we’re looking at F[xāˆ’c(t+Ī”t)] now, instead of F(xāˆ’ct). Now,Ā F[xāˆ’c(t+Ī”t)]Ā = F(xāˆ’ctāˆ’cĪ”t), so we’ll get a different valueĀ for our function—obviously! But it’s easy to see that we canĀ restore our wave function FĀ to its former value by alsoĀ adding some distanceĀ Ī”x =Ā cĪ”t to the argument. Indeed, if we do so, we get F[x+Ī”xāˆ’c(t+Ī”t)] = F(x+cĪ”t–ctāˆ’cĪ”t) = F(x–ct).Ā For example, if cĀ = 3 m/s, then 2 seconds of time correspond to (2 s)Ɨ(3 m/s) = 6 meters of distance.

The idea behind adding both some time Ī”tĀ as well as some distance Ī”x is that you’re traveling with the waveform itself, or with its phaseĀ as they say. So it’s like you’re riding on its crest or in its trough, or somewhere hanging on to it, so to speak. Hence, the speed of a wave is also referred to as itsĀ phase velocity, which we denote byĀ vpĀ =Ā c.Ā Now, let me make some remarks here.

First, there is theĀ directionĀ of travel. The pulses above travel in the positive x-direction, so that’s why we haveĀ x minusĀ ct in the argument. For a wave traveling in the negative x-direction, we’ll have a wave function y = F(x+ct). [And, yes, don’t be lazy, Vincent: please go through the Ī”x =Ā cĪ”t math once again to double-check that.]

The second thing you should note is that the speed of a regular periodic wave is equal to to the product of itsĀ wavelengthĀ and itsĀ frequency, so we write: vpĀ =Ā cĀ = λ·f, which we can also write as Ī» = c/f or f =Ā c/Ī». Now, you know we express the frequency in oscillations or cyclesĀ per second, i.e. in hertz: one hertzĀ is, quite simply, 1Ā sāˆ’1, so the unit of frequency is the reciprocal of the second. So the m/s and the Hz units in the fraction below give us a wavelengthĀ Ī» equal to Ī» =Ā (20 m/s)/(5/s) = 4 m. You’ll say that’s too simple but I just want to make sure you’ve got the basics right here.

Capture

The third thing is that, in physics, and in math, we’ll usually work with nice sinusoidal functions, i.e. sine or cosine functions. A sine and a cosine function are the same function but with aĀ phase differenceĀ of 90 degrees, so that’s Ļ€/2 radians.Ā That’s illustrated below: cosĪø = sin(Īø+Ļ€/2).

Circle_cos_sin

Now, when we converted time to distance by multiplying it with c, what we actually did was to ensure that the argument of our wavefunction F was expressed in one unitĀ only: the meter, so that’s theĀ distanceĀ unit in the international SI system of units. So that’s why we had to convert time to distance, so to speak.

The other option is to express all in seconds, so that’s in time units. So then we should measure distance in seconds, rather than meters, so to speak, and the corresponding argument is t–x/c, and our wave function would be written as y = G(t–x/c). Just go through the sameĀ Ī”x =Ā cĪ”t math once more: G[t+Ī”t–(x+Ī”x)/c] =Ā G(t+Ī”t–x/cāˆ’cĪ”t/c) =Ā G(t–x/c).

In short, we’re talking the same wave function here, soĀ F(xāˆ’ct) = G(tāˆ’x/c), but the argument of F is expressed in distance units, while the argument of G is expressed in time units. If you’d want to double-check what I am saying here, you can use the same 20 m/s wave example again: suppose the distance traveled is 100 m, so x = 100 m and x/cĀ = (100 m)/(20 m/s) = 5 seconds. It’s always important to check the units, and you can see they come out alright in both cases! šŸ™‚

Now, to go from F or G to our sine or cosine function, we need to do yet another conversion of units, as the argument of a sinusoidal function is some angleĀ Īø, not metersĀ orĀ seconds.Ā In physics, we refer toĀ Īø as theĀ phaseĀ of the wave function. So we need degrees or, more common now, radians, which I’ll explain in a moment.Ā Let me first jot it down:

y = sin(2Ļ€(x–ct)/Ī»)

So what are we doing here? What’s going on? Well… First, we divide x–ct by the wavelengthĀ Ī», so that’s the (x–ct)/Ī» in the argument of our sine function. So our ‘distance unit’ is no longer the meter but the wavelength of our wave, so we no longer measure inĀ meterĀ but in wavelengths. For example, if our argumentĀ x–ct was 20 m, and the wavelength of our wave is 4 m, we get (x–ct)/Ī» = 5 between the brackets. It’s just like comparing our length: ten years ago you were about half my size. Now you’re the same: one unit. šŸ™‚ When we’re saying that, we’re usingĀ myĀ length as the unit – and so that’s also yourĀ length unit now šŸ™‚Ā ā€“Ā rather than meters or centimeters.

Now I need to explain the 2Ļ€ factor, which is only slightly more difficult. Think about it: one wavelength corresponds to one full cycle, so that’s the full 360° of the circle below. In fact, we’ll express angles inĀ radians, and the two animations below illustrate what a radian really is: an angle of 1Ā rad defines an arc whose length, as measured on the circle, is equal to the radius of that circle. […] Oh! Please look at the animations as two separate things: they illustrate the same idea, but they’re notĀ synchronized, unfortunately! šŸ™‚
2pi-unrolledCircle_radians

So… I hope it all makes sense now: if we add one wavelength to the argument of our wave function, we should get the same value, and so it’s equivalent to adding 2Ļ€ to the argument of our sine function. Adding half a wavelength, or 35% of it, or a quarter, or twoĀ wavelengths, or e wavelengths, etcĀ is equivalent to adding π, orĀ 35%Ā·2Ļ€Ā ā‰ˆ 2.2, or 2Ļ€/4 = Ļ€/2, or 2Ā·2Ļ€ = 4Ļ€, or eĀ·2Ļ€, etc to it. So… Well… Think about it: to go from the argument of our wavefunction expressed as a number of wavelengthsĀ āˆ’ so that’s (x–ct)/Ī» – to the argument of our sine function, which is expressed in radians, we need to multiply by 2Ļ€.

[…] OK, Vincent. If it’s easier for you, you may want to think of the 1/Ī» and 2Ļ€ factors in the argument of theĀ sin(2Ļ€(x–ct)/Ī») function as scaling factors: you’d use a scaling factorĀ when you go from one measurement scale to another indeed. It’s like usingĀ vincentsĀ rather than meter. If oneĀ vincentĀ corresponds to 1.8 m, then we need to re-scale all lengths by dividingĀ them by 1.8 so as to express them inĀ vincents. Vincent ten year ago was 0.9 m, so that’s half a vincent: 0.9/1.8 = 0.5. šŸ™‚

[…] OK. […] Yes, you’re right: that’s rather stupid and makes nobody smile. Fine. You’re right: it’s time to move on to more complicated stuff. Now, read the following a couple of times. It’s my one and only message to you:

If there’s anything at all that you shouldĀ remember from all of the nonsense I am writing about in this physics blog, it’s thatĀ any periodic phenomenon, any motion really, can be analyzed by assuming that it is the sum of the motions of all the different modes of what we’re looking at, combined with appropriate amplitudes and phases.

It really is a most amazing thing—it’s something very deep and very beautiful connecting all of physics with math.

We often refer to these modes as harmonicsĀ and, in one of my posts on the topic, I explained how the wavelengths of the harmonics of a classical guitar string – it’s just an example – depended on the length of the string only. Indeed, if we denote the various harmonics by their harmonic number n = 1, 2, 3,… n,… and the length of the string by L, we have Ī»1Ā =Ā 2L = (1/1)Ā·2L, Ī»2Ā = L = (1/2)Ā·2L, Ī»3Ā =Ā (1/3)Ā·2L,… Ī»nĀ =Ā (1/n)Ā·2L. So they look like this:

620px-Harmonic_partials_on_strings

etcetera (1/8, 1/9,…,1/n,… 1/āˆž)

The diagram makes it look like it’s very obvious, but it’s an amazing fact: theĀ materialĀ of the string, or itsĀ tension, doesn’t matter. It’s just the length: simple geometry is all that matters! As I mentioned in my post on music and physics, this realization led to a somewhat misplaced fascination with harmonic ratios, which the Greeks thought could explain everything. For example, the Pythagorean model of the orbits of the planets would also refer to these harmonic ratios, and it took intellectual giants like Galileo and Copernicus to finally convince the PopeĀ that harmonic ratios are great, but that they cannot explainĀ everything. šŸ™‚ [Note:Ā When I say that theĀ materialĀ of the string, or itsĀ tension, doesn’t matter, I should correct myself: they do come into play whenĀ timeĀ becomes the variable. Also note that guitar strings are not the same length when strung on a guitar: the so-called bridge saddle is not in an exact right angle to the strings: this is a link to some close-up pictures of a bridge saddle on a guitar, just in case you don’t have a guitar at home to check.]

Now, I already explained the need to express the argument of a wave function in radians – because we’re talking periodicĀ functions and so we want toĀ use sinusoidals āˆ’ and how it’s just a matter of units really, and so how we can go from meter to wavelengths to radians. I also explained how we could do the same for seconds, i.e. for time. The key to convertingĀ distance units to time units, and vice versa, is the speed of the wave, or the phaseĀ velocity, which relates wavelength and frequency:Ā cĀ = λ·f. Now, as we have to express everything in radians anyway, we’ll usuallyĀ substitute the wavelength and frequency by theĀ wavenumberĀ and theĀ angular frequency so as to convert these quantities too to something expressed inĀ radians. Let me quickly explain how it works:

  1. The wavenumber k is equal to k = 2Ļ€/Ī», so it’s some number expressed inĀ radians per unit distance, i.e. radiansĀ per meter. In the example above, whereĀ Ī» was 4 m, we have k =Ā 2Ļ€/(4 m) = π/2 radiansĀ per meter. To put it differently, if our wave travels one meter, its phase Īø will change by π/2.
  2. Likewise, the angular frequency is ω = 2π·f = 2Ļ€/T. Using the same example once more, so assuming a frequency of 5 Hz, i.e. a periodĀ of one fifth of a second,Ā we have ω =Ā 2Ļ€/[(1/5)Ā·s] = 10Ļ€ per second. So the phase of our wave will change with 10 times π in one second. NowĀ thatĀ makes sense because, in one second, we have five cycles, and so that corresponds to 5 times 2Ļ€.

Note that our definition implies that Ī» = 2Ļ€/k, and that it’s also easy to figure out thatĀ our definition of ω, combined with the fĀ = c/λ relation, implies that ω = 2π·c/Ī» and, hence, that c = ω·λ/(2Ļ€) =Ā (ω·2Ļ€/k)/(2Ļ€) = ω/k.Ā OK. Let’s move on.

Using the definitions and explanations above, it’s now easy to see that we can re-write our y = sin(2Ļ€(x–ct)/Ī») as:

y = sin(2Ļ€(x–ct)/Ī») =Ā sin[2Ļ€(x–(ω/k)t)/(2Ļ€/k)] =Ā sin[(x–(ω/k)t)Ā·k)] =Ā sin(kx–ωt)

Remember, however, that we were talking some wave that was traveling in the positive x-direction. For theĀ negativeĀ x-direction, the equation becomes:

y = sin(2Ļ€(x+ct)/Ī») = sin(kx+ωt)

OK. That should be clear enough. Let’s go back to our guitar string. We can go from Ī» to k by noting that Ī» = 2L and, hence, we get the following for all of the various modes:

k = k1Ā = 2π·1/(2L) = Ļ€/L, k2Ā = 2π·2/(2L) = 2k, k3Ā =Ā 2π·3/(2L) = 3k,,… knĀ =Ā 2π·3/(2L) = nk,…

That gives us our grand result, and that’s that we can write some very complicated waveform ĪØ(x) as the sum of an infinite number of simple sinusoids, so we have:

ĪØ(x) =Ā a1sin(kx) + a2sin(2kx) + a3sin(3kx) + … + ansin(nkx) + … =Ā āˆ‘Ā ansin(nkx)

The equation above assumes we’re looking at the oscillation at some fixed point in time. If we’d be looking at the oscillation at some fixed point in space, we’d write:

Φ(t) = a1sin(ωt) + a2sin(2ωt) + a3sin(3ωt) + … + ansin(nωt) + … =Ā āˆ‘Ā ansin(nωt)

Of course, to represent some very complicated oscillation on our guitar string,Ā we can and should combine someĀ ĪØ(x) as well as some Φ(t) function, butĀ how do we do that,Ā exactly? Well… We’ll obviously need both theĀ sin(kx–ωt) as well as those sin(kx+ωt) functions, as I’ll explain in a moment. However, let me first make another small digression, so as to complete your knowledge of wave mechanics. šŸ™‚

We look at a wave as something that’s traveling through space and time at the same time. In that regard, I told you that the speedĀ of the wave is its so-calledĀ phase velocity, which we denoted as vpĀ = cĀ and which, as I explained above, is equal toĀ vpĀ =Ā cĀ = λ·fĀ = (2Ļ€/k)Ā·(ω/2Ļ€) = ω/k. The animation below (credit for it must go to Wikipedia—and sorry I forget to acknowledge the same source for the illustrations above) illustrates the principle: the speed of travel of the red dot is the phase velocity. But you can see that what’s going on here is somewhat more complicated: we have a series of wave packetsĀ traveling through space and time here, and so that’s where the concept of the so-calledĀ groupĀ velocity comes in: it’s the speed of travel of the green dot.

Wave_group

Now, look at the animation below. What’s going on here? The wave packet (or theĀ groupĀ or theĀ envelopeĀ of the wave—whatever you want to call it) moves to the right, but the phase goes to the left, as the peaks and troughs move leftward indeed. Huh?Ā How is that possible? And where is this wave going? Left or right? Can we still associate someĀ directionĀ with the wave here?Ā It looks like it’s traveling in both directions at the same time!

Wave_opposite-group-phase-velocity

The wave actually does travel in both directions at the same time. Well… Sort of. TheĀ point is actually quite subtle. When I started this post by writing that the pulses were ‘obviously’ traveling in the positive x-direction… Well… That’s actually notĀ so obvious. What is it that is traveling really? Think about an oscillating guitar string: nothing travels left or right really. Each point on the string just moves up and down. Likewise, if our repeated pulse is some water wave, then the water just stays where it is: it just moves up and down. Likewise, if we shake up some rope, the rope is not going anywhere: we just started some motionĀ that is traveling down the rope.Ā In other words, the phase velocity is just a mathematical concept. The peaks and troughs that seem to be traveling are just mathematical points that are ‘traveling’ left or right.

What about the group velocity? Is that a mathematical notion too? It is. The wave packet is often referred to as the envelopeĀ of the wave curves, for obviously reasons: they’re enveloped indeed. Well… Sort of. šŸ™‚ However, while both the phase and group velocity are velocities of mathematical constructs, it’s obvious that, if we’re looking at wave packets, the group velocity would be of more interest to us than the phase velocity. Think of those repeated pulses as real water waves, for example: while the water stays where it is (as mentioned, the water molecules just go up and down—more or less, at least), we’d surely be interested to know how fast these waves are ‘moving’, and that’s given by the groupĀ velocity, not the phase velocity. Still, having said that, the group velocity is as ‘unreal’ as the phase velocity: both are mathematical concepts. The only thing that’s ‘real’ is the up and down movement. Nothing travelsĀ in reality. Now, I shouldn’t digress too much here, but that’s why there’s no limit on the phase velocity: it can exceed the speed of light. In fact, in quantum mechanics, some real-life particle āˆ’ like an electron, for instance – will be represented by a complex-valuedĀ wave function, and there’s no reason to put some limit on the phase velocity. In contrast, theĀ groupĀ velocity will actually be the speed of the electron itself, andĀ that speed can, obviously, approach the speed of light – in particle accelerators, for example – but it can never exceed it. [If you’re smart, and you are, you’ll wonder: what about photons? Well…The classical and quantum-mechanical view of an electromagnetic wave are surelyĀ notĀ the same, but they do have a lot in common: both photons and electromagnetic radiation travel at the speedĀ c. Photons can do so because their rest mass is zero. But I can’t go into any more detail here, otherwise this thing will become way too long.]

OK. Let me get back to the issue at hand. So I’ll now revert to the simpler situation we’re looking at here, and so that’s theseĀ harmonicĀ waves, whose form is a simple sinusoidal indeed. The animation below (and, yes, it’s also from Wikipedia) is the one that’s relevant for this situation. You need to study it for a while to understand what’s going on. As you can see, the green wave travels to the right, the blue one travels to the left, and the red wave function is the sum of both.

Superpositionprinciple

Of course, after all that I wrote above, I should use quotation marks and write ‘travel’ instead of travel, so as to indicate there’s nothing traveling really, except for those mathematical points, but then no one does that, and so I won’t do it either. Just make sure you always think twice when reading stuff like this! Back to the lesson: what’s going on here?

As I explained, the argument of a wave traveling towards the negativeĀ x-direction will be x+ct. Conversely, the argument of a wave traveling in theĀ positiveĀ x-direction will be x–ct. Now, our guitar string is going nowhere, obviously: it’s like the red wave function above. It’s a so-called standing wave.Ā The red wave function has nodes, i.e. points where there is no motion—no displacement at all! Between the nodes, every point moves up and down sinusoidally, but the pattern of motion stays fixed in space.Ā So that’s the kind of wave function we want, and the animation shows us how we can get it.

Indeed, there’s a funnyĀ thing with fixed strings:Ā when a wave reaches the clamped end of a string, it will be reflected with a change in sign, as illustrated below: we’ve got that F(x+ct) wave coming in, and then it goes back indeed, but with the sign reversed.

Capture

The illustration above speaks for itself but, of course, once again I need to warn you about the use of sentences like ‘the wave reaches the end of the string’ and/or ‘the wave gets reflected back’. You know what it really means now: it’s some movementĀ that travels through space. […] In any case, let’s get back to the lesson once more:Ā how do we analyze that?

Easy: the red wave function is the sum of two waves: one traveling to the right, and one traveling to the left. We’ll call these componentĀ waves F and G respectively, so we have y = F(x, t) + G(x, t). Let’s go for it.

Let’s first assume the string isĀ notĀ held anywhere, so that we have an infinite string along which waves can travel in either direction. In fact, the most general functional form to capture the fact that a waveform can travel in any direction is to write the displacement y as the sum ofĀ twoĀ functions: one wave traveling one way (which we’ll denote by F, indeed), and the other wave (which, yes, we’ll denote by G) traveling the other way. From the illustration above, it’s obvious that the F wave is traveling towards the negativeĀ x-direction and, hence, its argument will be x+ct. Conversely, the G wave travels in theĀ positiveĀ x-direction, so its argument is x–ct.Ā So we write:

y = F(x, t) + G(x, t) = F(x+ct) + G(x–ct)

So… Well… We knowĀ that the string is actually not infinite, but that it’s fixed to two points. Hence, y is equal to zero there: y = 0. Now let’s choose the origin of our x-axis at the fixed end so as to simplify the analysis. Hence, where y is zero, x is also zero. Now, at x = 0, our general solution above for the infinite string becomes Ā y = F(ct) + G(āˆ’ct) = 0, for all values of t. Of course, that means G(āˆ’ct) must be equal to –F(ct). Now, that equality is there for all values of t. So it’s there for all values of ct and āˆ’ct. In short, that equality is valid for whatever value of the argument ofĀ G and –F. As Feynman puts it: ā€œGĀ of anything must be –FĀ of minus that same thing.ā€ Now, the ā€˜anything’ in G is its argument: x – ct, so ā€˜minus that same thing’ is –(x–ct) =Ā āˆ’x+ct. Therefore, our equation becomes:

y = F(x+ct) āˆ’Ā F(āˆ’x+ct)

So that’s what’s depicted in the diagram above: the F(x+ct) wave ā€˜vanishes’ behind the wall as the āˆ’Ā F(āˆ’x+ct) wave comes out of it.Ā Now, of course, so as to make sure our guitar string doesn’t stop its vibration after being plucked, we need to ensure F is a periodic function, like aĀ sin(kx+ωt) function. šŸ™‚ Why? Well… If this F and G function would simply disappear and ‘serve’ only once, so to speak, then we only have one oscillation and that’s it! So the waves need to continue and so that’s why it needs to be periodic.

OK. Can we just takeĀ sin(kx+ωt) and āˆ’sin(āˆ’kx+ωt) and add both? It makes sense, doesn’t it? Indeed, āˆ’sinα = sin(āˆ’Ī±) and, therefore, āˆ’sin(āˆ’kx+ωt) = sin(kxāˆ’Ļ‰t). Hence,Ā y = F(x+ct) āˆ’Ā F(āˆ’x+ct) would be equal to:

y =Ā sin(kx+ωt) + sin(kx–ωt) = sin(2Ļ€(x+ct)/Ī») + sin(2Ļ€(xāˆ’ct)/Ī»)

Done! Let’s use specific values for k and ω now. For the first harmonic, we know that k =Ā 2Ļ€/2L = Ļ€/L. What about ω? Hmm… That depends on the wave velocity and, therefore, that actually doesĀ dependĀ on the material and/or the tension of the string! The only thing we can say is that ω = cĀ·k, so ω =Ā cĀ·2Ļ€/Ī» = cĀ·Ļ€/L. So we get:

sin(kx+ωt) = sin(π·x/L + π·cĀ·t/L) =Ā sin[(Ļ€/L)Ā·(x+ct)]

But this is our F function only. The whole oscillation isĀ y = F(x+ct) āˆ’Ā F(āˆ’x+ct), and āˆ’Ā F(āˆ’x+ct) is equal to:

 –sin[(Ļ€/L)Ā·(āˆ’x+ct)] = –sin(āˆ’Ļ€Ā·x/L+π·cĀ·t/L) =Ā āˆ’sin(āˆ’kx+ωt) = sin(kx–ωt) =Ā sin[(Ļ€/L)Ā·(x–ct)]

So, yes, we should add both functions to get:

y = sin[Ļ€(x+ct)/L] + sin[Ļ€(xāˆ’ct)/L]

Now, we can, of course, apply our trigonometric formulas for the addition of angles, which say that sin(α+β) =Ā sinαcosβ + sinβcosα and sin(α–β) =Ā sinαcosβ – sinβcosα. Hence, y =Ā sin(kx+ωt) +Ā sin(kx–ωt) is equal toĀ sin(kx)cos(ωt) + sin(ωt)cos(kx) + sin(kx)cos(ωt) – sin(ωt)cos(kx) = 2sin(kx)cos(ωt). Now, that’s a very interesting result, so let’s give it some more prominence by writing it in boldface:

y =Ā sin(kx+ωt) +Ā sin(kx–ωt) =Ā 2sin(kx)cos(ωt) =Ā 2sin(π·x/L)cos(π·cĀ·t/L)

TheĀ sin(π·x/L) factor gives us the nodes in space. Indeed,Ā sin(π·x/L) = 0 ifĀ x is equal to 0 or L (values of x outside of the [0, L] interval are obviously notĀ relevant here). Now, the other factorĀ cos(π·cĀ·t/L) can be re-written cos(2π·cĀ·t/Ī») = cos(2π·fĀ·t) = cos(2π·t/T), with T the periodĀ T = 1/fĀ = Ī»/c, so the amplitude reaches a maximum (+1 orĀ āˆ’1 or, including the factor 2, +2 or āˆ’2) if 2π·t/T is equal to a multiple of Ļ€, so that’s ifĀ t = nĀ·T/2 with n = 0, 1, 2, etc. In our example above, for f = 5 Hz, that means the amplitude reaches a maximum (+2 or āˆ’2) every tenthĀ of a second.

The analysis for the other modes is as easy, and I’ll leave it you, Vincent,Ā as an exercise, to work it all out and send me the y = 2Ā·sin[something]Ā·cos[something else] formula (with the ‘something’ and ‘something else’ writtenĀ in terms of L and c, of course) for the higher harmonics.Ā šŸ™‚

[…] You’ll say: what’s the point, daddy? Well… Look at that animation again: isn’t it great we can analyze any standing wave, or any harmonic indeed, as the sum ofĀ twoĀ component waves with the same wavelength and frequency but ‘traveling’ in opposite directions?Ā 

Superpositionprinciple

Yes, Vincent. I can hear you sigh: “Daddy, I really do notĀ see why I should be interested in this.”

Well… Your call… What can I say? Maybe one day you will. In fact, if you’re going to go for engineering studies, you’ll have to. šŸ™‚

To conclude this post, I’ll insert one more illustration. Now that you know whatĀ modesĀ are, you can start thinking about those more complicatedĀ ĪØ and Φ functions. The illustration below shows how the first and second mode of our guitar string combine to give us some composite wave traveling up and down the very same string.

composite wave

Think about it. We haveĀ oneĀ physical phenomenon here: at every point in time, the string is somewhere, butĀ where exactly, depends on the mathematical shape of its components. If this doesn’t illustrate the beauty of Nature, the fact that, behind every simple physicalĀ phenomenon āˆ’ most of which are some sort of oscillation indeed āˆ’ we have some marvelous mathematicalĀ structure, then… Well… Then I don’t know how to explain why IĀ am absolutely fascinated by this stuff.

Addendum 1: On actual waves

My examples of waves above were all examples of so-called transverseĀ waves, i.e. oscillations at a right angle to the direction of the wave. The other type of wave is longitudinal. I mentioned sound waves above, but they are essentiallyĀ longitudinal. So thereĀ  the displacement of the medium is in the same direction of the wave, as illustrated below.Onde_compression_impulsion_1d_30_petit

Real-life waves, like water waves, may be neither of the two. The illustration below shows how water molecules actually move as a wave passes. They move in little circles, with a systemic phase shift from circle to circle.

Water waves

Why is this so? I’ll let Feynman answer, as he also provided the illustration above:

“Although the water at a given place is alternately trough or hill, it cannot simply be moving up and down, by the conservation of water. That is, if it goes down, where is the water going to go? The water is essentially incompressible. The speed of compression of waves—that is, sound in the water—is much, much higher, and we are not considering that now. Since water is incompressible on this scale, as a hill comes down the water must move away from the region. What actually happens is that particles of water near the surface move approximately in circles. When smooth swells are coming, a person floating in a tire can look at a nearby object and see it going in a circle. So it is a mixture of longitudinal and transverse, to add to the confusion. At greater depths in the water the motions are smaller circles until, reasonably far down, there is nothing left of the motion.”

So… There you go… šŸ™‚

Addendum 2: On non-periodic waves, i.e. pulses

A waveform isĀ notĀ necessarily periodic. The pulse we looked at could, perhaps,Ā not repeat itself. It is notĀ possible, then, to describe its wavelength. However, it’s still a wave and, hence, its functional form would still be some y = F(xāˆ’ct) or y = F(x+ct) form, depending on its direction of travel.

The example below also comes out of Feynman’s Lectures: electromagnetic radiation is caused by someĀ acceleratingĀ electric charge – an electron, usually, because its mass is small and, hence, it’s much easier to move than a proton šŸ™‚ – and then the electric field travels out in space.Ā So the two diagrams below show (i) the acceleration (a) as a function of time (t) and (ii) the electric field strength (E) as a function of the distance (r). [To be fully precise, I should add he ignores the 1/r variation, but that’s a fine point which doesn’t matter much here.]

EM 1EM 2 Ā Ā Ā 

He basically uses this illustration to explain why we can use a y =Ā G(t–x/c) functional form to describe a wave. The point is: he actually talks about oneĀ pulse only here. So theĀ F(x±ct) or G(t±x/c) or sin(kx±ωt) form hasĀ nothingĀ to do with whether or not we’re looking at a periodic or non-periodic waveform. The gist of the matter is that we’ve got something movingĀ through space, and it doesn’t matter whether it’s periodic or not: theĀ periodicity or non-periodicity,Ā of a wave has nothing to do with the x±ct, t±x/cĀ or kx±ωt shape of the argument of our wave function. The functional form of our argument is just the result of what I said about traveling along with our wave.

So what is it about periodicity then? Well… If periodicity kicks it, you’ll talk sinusoidal functions, and so the circle will be needed once more. šŸ™‚Ā Ā Ā 

Now, I mentioned we cannotĀ associate any particular wavelength with such non-periodic wave. Having said that, it’s still possible to analyze this pulseĀ as a sum ofĀ sinusoids through a mathematical procedure which is referred to as theĀ Fourier transform. If you’re going for engineer, you’ll need to learn how to master this technique. As for now, however, you can just have a look at the Wikipedia article on it. šŸ™‚

Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/
Some content on this page was disabled on June 16, 2020 as a result of a DMCA takedown notice from The California Institute of Technology. You can learn more about the DMCA here:

https://wordpress.com/support/copyright-and-the-dmca/