We're sorry but this page doesn't work properly without JavaScript enabled. Please enable it to continue.
Feedback

4. Spin One-half, Bras, Kets, and Operators

00:00

Formal Metadata

Title
4. Spin One-half, Bras, Kets, and Operators
Title of Series
Number of Parts
25
Author
License
CC Attribution - NonCommercial - ShareAlike 4.0 International:
You are free to use, adapt and copy, distribute and transmit the work or content in adapted or unchanged form for any legal and non-commercial purpose as long as the work is attributed to the author in the manner specified by the author or licensor and the work or content is shared also in adapted form only under the conditions of this
Identifiers
Publisher
Release Date
Language

Content Metadata

Subject Area
Genre
Abstract
In this lecture, the professor talked about spin one-half states and operators, properties of Pauli matrices and index notation, spin states in arbitrary direction, etc.
ElectronRail transport operationsSchubvektorsteuerungSilveringAngeregter ZustandSpin (physics)Screen printingMachineMeasurementElectronic componentSpare partBasis (linear algebra)Direct currentHourTypesettingPackaging and labelingGaussian beamMagnetic momentQuality (business)Hose couplingMaterialAtomismGround stationLastParticleQuantumCartridge (firearms)Lecture/Conference
PagerGalaxyAngeregter ZustandSchubvektorsteuerungParticleBasis (linear algebra)Electronic componentBracket clockFACTS (newspaper)Spin (physics)Avro Canada CF-105 ArrowProzessleittechnikSteckverbinderTypesettingLecture/Conference
H-alphaAngeregter ZustandStarBeta particleSchubvektorsteuerungNanotechnologyBracket clockMatrix (printing)Spin (physics)Electronic componentWatercraft rowingRail transport operationsBasis (linear algebra)Cartridge (firearms)MachineAvro Canada CF-105 ArrowMultiplizitätCombined cycleRoots-type superchargerAmplitudeLecture/Conference
Angeregter ZustandRail transport operationsLastDirect currentElectronic componentMatrix (printing)Spin (physics)SchubvektorsteuerungHose couplingFlugbahnMeasurementNanotechnologyFACTS (newspaper)Cartridge (firearms)SigmaAnalog signalWind waveGas turbineLambda baryonProzessleittechnikCombined cycleAmplitudeAmateur radio repeaterMachineRefractive indexLecture/Conference
Nuclear fuelAtmosphere of EarthBasis (linear algebra)SchubvektorsteuerungMatrix (printing)Scale (map)TypesettingMultiplizitätRail transport operationsSpare partPhase (matter)Angeregter ZustandDirect currentLambda baryonAngle of attackOrder and disorder (physics)Spin (physics)Hot workingVideoElectronic componentProgressive lensLecture/Conference
Lambda baryonSigmaMatrix (printing)Roots-type superchargerRoll formingPhase (matter)Electronic componentPower (physics)Spare partWatercraft rowingDrehmasseAngeregter ZustandBending (metalworking)CogenerationConstraint (mathematics)Basis (linear algebra)Linear motorRail transport operationsLecture/Conference
Angeregter ZustandMatrix (printing)Lambda baryonRail transport operationsRoots-type superchargerSchubvektorsteuerungFACTS (newspaper)Direct currentMeasurementAmplitudePlane (tool)Combined cycleYElectronic componentLecture/Conference
SchubvektorsteuerungLastTurningMatrix (printing)SigmaRail transport operationsHose couplingNanotechnologyFlugbahnDirect currentLambda baryonSpin (physics)MeasurementAngeregter ZustandWeather frontElectronic componentHot workingLecture/Conference
Phase (matter)Angeregter ZustandRail transport operationsAngle of attackVideoOrder and disorder (physics)Electronic componentDirect currentMatrix (printing)SchubvektorsteuerungBasis (linear algebra)Spin (physics)Lecture/Conference
Transcript: English(auto-generated)
The following content is provided under a Creative Commons license. Your support will help MIT OpenCourseWare continue to offer high quality educational resources for free. To make a donation, or to view additional materials from hundreds of MIT courses, visit MIT OpenCourseWare at ocw.mit.edu.
PROFESSOR ERIK DEMAINE-MARTINI Last time we spoke about the Stern-Gerlach experiment, and how you could have a sequence of Stern-Gerlach boxes that allow you to understand the type of states
and properties of the physics having to do with spin 1 half. So the key thing in the Stern-Gerlach machine was that a beam of silver atoms, each of which is really like an electron with a magnetic moment,
was placed in an inhomogeneous, strong magnetic field. And that would classically mean that you would get a deflection proportional to the z component of the magnetic moment. What was a surprise that by the time you put the screen on the right side,
it really split into two different beams, as if the magnetic moments could either be all the way pointing in the z direction or all the way down, pointing opposite to the z direction, and nothing in between. A very surprising result.
So after looking at a few of those boxes, we decided that we would try to model the spin 1 half particle as a two-dimensional complex vector space. What is the two-dimensional complex vector space?
It's the possible space of states of a spin 1 half particle. So our task today is to go in detail into that and set up the whole machinery of spin 1 half. So we will do so, even though we haven't quite yet
discussed all the important concepts of linear algebra that we're going to need. So today I'm going to assume that at least you have some vague notions of linear algebra reasonably well understood. And if you don't, well, take them on faith today.
We're going to go through them slowly in the next couple of lectures. And then, as you will reread this material, it will make more sense. So what did we have? We said that the spin states, or the possible states
of this silver atom, that really correspond to an electron could be described by states z comma plus and z colon minus.
So these are the two states. This state, we say, corresponds to an angular momentum, Sz hat, Sz, let me put it like that, of h bar over 2.
And this corresponds to Sz equal minus h bar over 2. And those are our two states. The z label indicates that we've passed, presumably, these atoms through a filter in the z
direction so that we know for certain we're talking about the z component of angular momentum of this state. And it is positive, and the value is this. Here, again, we have the label z to remind us that we're talking about states that
have been organized using the z component of angular momentum. You could ask whether this state has some angular momentum in the spin angular momentum in the x direction or in the y direction, and we will be able to answer that question in an hour from now.
So mathematically, we say that this statement, that this state has Sz equal h bar over 2 means that there is an operator, an operator Sz hat,
hat for operators. And this operator, we say, acts on this state
to give h bar over 2 times this state. So when we have a measurement in quantum mechanics, we end up talking about operators. So this case is no exception. We think of the operator Sz that acts on this state
and gives h bar over 2. And that same operator Sz acts on the other state and gives you minus h bar over 2 times this state.
You see, an operator on a state does give a state. So in this equation, we have a state on the right. And the nice thing is that the same state appears on the right. When that happens, you say that the state is an eigenstate of the operator. And therefore, the states z plus minus
are eigenstates of the operator Sz with eigenvalues, the number that appears here, equal to plus minus h over 2. So the relevant physical assumption here
is the following, that these two states, in a sense, suffice. Now, what does that mean? We could do the experiment again with a Stern-Gerlach machine that is along the x-axis and say, oh, now we've got states x plus and x minus.
And we should add them there. They're also part of the possible states of the system. Kind of. They're parts of the possible states of the system. They are possible states of the system. But we shouldn't add them to these ones.
These will be thought as basis states, just like any vector is a superposition of a number times the x-unit vector plus a number times the y-unit vector and a number times the z-unit vector. We are going to postulate or try to construct the theory of spin based on the idea
that all possible spin states of an electron are obtained by suitable linear superposition of these two vectors. So in fact, what we're going to say is that these two vectors are the basis
of a two-dimensional vector space such that every possible state is a linear superposition. So psi being any possible spin state
can be written as some constant c1 times z plus c2 times z minus, where this constant c1 and c2
belong to the complex numbers. And by this, we mean that if any possible states is a superposition like that, the set of all possible states are the general vectors in a two-dimensional complex vector
space. Complex vector space because the coefficients are complex, and two-dimensional because there's two basis vectors. Now, this doesn't quite look like a vector. It looks like this thing's called kets.
But kets are really vectors. And we're going to make the correspondence very clear. So this can be called the first basis state
and the second basis state. And I want you to realize that the fact that we're talking about the complex vector space really means these coefficients are complex. There's no claim that the vector is complex in any sense
or this one. They're just vectors. And this is a vector. And it's not like we say, oh, this vector is complex. No, a complex vector space, we think of a set of vectors. And then we're allowed to multiply them
by complex numbers. OK. OK. So we have this. And this way of thinking of the vectors is quite all right.
But we want to be more concrete. For that, we're going to use what is called a representation. So I will use the word representation to mean some way of exhibiting a vector or a state in a more concrete way, as something that any one of us
would call a vector. So as a matter of notation, this being the first basis state is sometimes written as a ket with a 1, like that.
And this being the second basis state is sometimes written this way. But here is the real issue of what we are calling a representation. If this is a two-dimensional vector space, you're accustomed, a three-dimensional vector
space, what are vectors? They're triplets of numbers, three numbers. That's a vector. Column vector is perhaps easier to think about them. So column vector. So here's what we're going to say. We have the state z plus. It's also called 1.
It's just a name. But we're going to represent it as a column vector. And as a column vector, I'm going to represent it as the column vector 1, 0.
And this is why I put this double arrow. I'm not saying it's the same thing, although really it is. It's just a way of thinking about it as some vector in what we would call, canonically, a vector space. Yes?
So that would be based on the actual physical process that's going on. Or what is their connection to the actual physical prime that they're representing here? Well, we'll see it in a second. It will become a little clearer. But this is like saying, I have
a two-dimensional vector space. So I'm going to think of the first state as this vector. But how do I write this vector? Well, it's the vector Ex. Well, if I would write them in components, I would say a vector, I can put two numbers here, A and B.
And this is the A component and B component. So here it is. Ex would be 1, 0. And Ey would be 0, 1. If I have this notation, that the point AB is represented by A and B as a column vector.
So at this moment, it's just a way of associating a vector in the two-dimensional, canonical vector space as just a column here. So the other state, minus, is also called 2, will be represented by 0, 1.
And therefore, this state, psi, which is c1 z plus plus c2 z
minus, will be represented as c1 times the first vector plus c2 times the second vector. Or multiplying in c1, c2.
So this state can be written as a linear superposition of these two basis vectors in this way. You can write it this way. You want to save some writing. You could write them with 1 and 2.
But as a vector, it is represented by a column vector with two components. That's our state. Now in doing this, I want to emphasize we're introducing the physical assumption that this will be enough to describe all possible spin
states, which is far from obvious at this stage. Nevertheless, let's use some ideas from the Stern-Gerlach experiment. We did one example of a box that filtered the plus z states
and then put it against another z machine. And then all the states went through the op, which is to say the plus states have no amplitude, no probability to be in the minus states. They all went and go through the plus.
So we're going to introduce now the physical translation of this fact as saying that these states are orthogonal to each other. So this will require the whole framework in detail of brass and kets to say really precisely.
But we're going to do that now and explain the minimum necessary for you to understand it. But we'll come back to it later. So this physical statement will be stated as z minus with z plus.
The overlap, the bracket of this, is 0. The fact that all particles went through and went out through the plus output will state it as, well, these states are well normalized.
So z plus z plus is 1. Similarly, you could have blocked the other input and you would have concluded that the minus state is orthogonal to the plus. So we also say that these two are orthogonal
and the minus states are well normalized. Now, here we had to write four equations. And the notation 1 and 2 comes handy
because we can summarize all these statements by the equation ij equals delta ij. Look, if this equation is 2 with 1 equals 0,
the bra 2, ket 1, this is 1 with 1 is equal to 1. Here is 1 with 2 is equal to 0 and 2 with 2 is equal to 1.
So this is exactly what we have here. Now, I didn't define for you these so-called bras. So by completeness, I will define them now. So given, and the way I will define them is as follows.
I will say that while for the one vector basis state you associate at 1, 0, you will associate to one bra
the row vector 1, 0. I sometimes tend to write equal, but equal is all right, but a little clearer to say that there's arrows here. So we're going to associate to 1, 1, 0.
We did it before. But now to the bra, we think of the row vector like this. Similarly, I can do the following the same with 2. 2 was the vector 0, 1 as a column vector. So 2 as a bra, we will think of it as the row vector 0, 1.
We're going to do this now a little more generally. So suppose you have a state alpha, which is alpha 1, 1 plus alpha 2, 2.
Well, to this you would associate the column vector alpha 1, alpha 2. Suppose you have a beta state beta 1, 1 plus beta 2, 2. You would associate beta 1, beta 2 as the representations.
Now here comes a definition for which this is just a special case. And it's a definition of the general bra. So the general bra here, alpha, is
defined to be alpha 1 star bra of the first plus alpha 2 star bra of the second. So this is alpha 1 star times the first bra,
which we think of it as 1, 0, plus alpha 2 star times the second bra, which is 0, 1. So this whole thing is represented by alpha 1 star,
alpha 2 star. So here we had the column vector representation
of a state. And the bra is the row vector representation of a state in which this is constructed with complex conjugation. Now this kind of definitions will be discussed in more detail and more axiomatically very soon
so that you see where they're going. But the intuition that you're going to get from this is quite valuable. So what is the bracket? Alpha beta is the so-called bracket. And this is a number.
And the reason for complex conjugation is ultimately that when these two things are the same, it should give a positive number. It's like the length squared. So that's the reason for complex conjugation eventually. But for now, you're supposed to get a number from here.
And the reasonable way to get a number, which is a definition, is that you get a number by matrix multiplication of the representative. So you take the representative of alpha,
which is alpha 1 star, alpha 2 star, and do the matrix product with the representative of beta, which is beta 1, beta 2. And that's alpha 1 star beta 1 plus alpha 2 star beta 2.
And that's the number called the inner product, or bracket product. And these are the true meanings of relations of this kind. If you're given arbitrary states, you compute the inner product this way.
And vectors that satisfy this are called orthonormal, because they're orthogonal and normal with respect to each other in the sense of the bra and ket. So this definition, as you can see,
is also consistent with what you have up there. And you can check it. You take i with j, 1, say, with 2, like this.
You do the inner product, and you get 0. And similarly for all the other states. So let's then complete the issue of representations. We had representations of the states
as column vectors, two by two column vectors, or row vectors. Now let's talk about this operator we started with. If this is an operator acting on states, now I want to think of its representation, which
would be the way it acts on these two component vectors. So it must be a two by two matrix, because only a two by two matrix acts naturally on two component vectors. So here is the claim that we have, claim that Sz hat is represented,
but we'll just put equal, by this matrix. You see, it was an operator.
We never talked about matrices. But once we start talking of the basis vectors as column vectors, then you can ask if this is correct. So for example, I'm supposed to find that Sz hat acting on the state 1
is supposed to be h bar over 2 times the state 1. Is it true? Then you say, oh, let's put the representation. h bar over 2, 1 minus 1, 0, 0. State 1, what's its representation?
1, 0. OK, let's act on it. So this gives me h bar over 2. I do the first product, I get a 1. I do the second product, I get a 0. Oh, that seems right, because this is h over 2 times the representation of the state 1.
And if I check this, and as well that Sz on 2 is equal to minus h bar over 2, 2, which can also be checked, I need check no more.
Because it suffices that this operator does what it's supposed to do of the basis vectors. And it will do what it's supposed to do on arbitrary vectors. So we're done. This is the operator Sx. And we seem to have put together a lot
of these ideas of the experiment into a mathematical framework. But we're not through, because we have this question. So what if you align the operate the machine along x? What are the possible spin states along the x direction?
How do you know that the spin state that points along x can be described in this vector space? How do I know there exists numbers c1, c2, so that this linear combination is a spin state that points along x?
Well, at this moment, you have really to invent something. And the process of invention is never a very linear one. It's you use analogies. You use whatever you can to invent what you need.
So given that that's a possibility, we could follow what Feynman does in his Feynman lectures of discussing how to begin rotating Stern-Gerlach machines and doing all kinds of things. It's an interesting argument. And it's a little hard to follow, a little tedious
at points. And we're going to follow a different route. I'm going to assume that you remember a little about angular momentum. And I think you do remember this much. I want to say, well, this is spin angular momentum.
So well, let's compare with orbital angular momentum and see where we are. You see, another way of asking the question would be, well, what are the operators Sx and Sy?
Where do I get them? Well, the reason I want to bring in the angular momentum is because there you have Lz. But you also have Lx and Ly. So angular momentum had Lz, just like we had here, but also Lx and Ly.
Now these spin things look a lot more mysterious, a lot more basic, because like Lz was xpy minus ypx.
So you knew how this operator acts on wave functions. You know, it multiplies by y, takes an x derivative, or it's a d d phi.
It has a nice thing. But Sz, on the other hand, there's no x. There's no derivatives. It's a different space. It's working in a totally different space, in a space of a two-dimensional complex vector space, a column vectors with two numbers.
That's where it acts. I'm sorry. There's no d dx, nothing familiar about it. But that's what we have been handed. So this thing acts on wave functions and does natural things. Well, the other one acts on column vectors, two-by-two, two-component column vectors.
And that's all right. But we also know that Lz is Hermitian. And that was good, because it actually meant that it's a good observable. You can measure it. Is Sz Hermitian? Well, yes, it is.
Hermiticity of a matrix, as we will discuss in a lot of detail, maybe more than you want, means you can transpose it, complex conjugate it, and you get the same matrix. Well, that matrix is Hermitian, so that's nice. That maybe is important.
So what other operators we had, Lx and Ly, and if we think of Lx as L1, Ly as L2, and Lz as L3, you had a basic commutation relation, Li with Lj,
was equal to i epsilon ijk Lk hat ih bar. And this was called the algebra of angular momentum.
These three operators satisfy these identities. i and j are here. k is supposed to be summed over. Repeated indices are summed from 1, 2, and 3. And epsilon ijk is totally anti-symmetric, with epsilon 1, 2, 3 equal to plus 1.
You may or may not know this epsilon. You will get some practice on that very soon. Now, for all intents and purposes, we might as well write the explicit formulas, which read Lx Ly equal ih bar Lz, Ly Lz equal ih bar Lx,
and Lz Lx, their hats all over, equal ih bar Ly.
So we had this for orbital angular momentum, or for angular momentum in general.
So what we're going to do now is going to try to figure out what Sx, what are Sx and Sy, by trying to find complete analogy. We're going to declare that S is
going to be angular momentum. So we're going to want that Sx with Sy will be ih bar Sz. Sy with Sz will be ih bar Sx.
And finally, Sz with Sx is ih bar Sy. And we're going to try that these things be Hermitian, Sx
and Sy. So let me break for a second and ask if there are questions. We're aiming to complete the theory by taking S to be angular momentum and see what we get.
Can we invent operators Sx and Sy that will do the right thing? Yes? What's the name for the epsilon ijk? I know there's a special name from Cevita. What's the name? Levi-Civita tensor. Levi-Civita tensor, that's right.
Levi-Civita tensor can be used for cross products. It's very useful for cross products. It's a really useful tensor. Other questions? More questions about what we're going to try to do or this so far?
Yes? So when you use the term representation, does that have a, is that like the technical, mathematical term of representation like in algebra? Yes, it's representation of operators in vector spaces. So we've used the canonical vector space with column
vectors represented by entries 1 and numbers. And then the operators become matrices. So whenever an operator is viewed as a matrix, we think of it as a representation. Other questions? Yes? Will we talk about later why we
can make an analogy between L and S or? Well, you see, this is a very strong analogy. But there will be big differences
from orbital angular momentum and spin angular momentum. And basically, having to do with the fact that the eigenvalues of these operators are plus minus h bar over 2. And in the orbital case, they tend to be plus minus integer values of h bar. So this is a very deep statement
about the algebra of these operators that still allows the physics of them to be quite different. But this is probably the only algebra that makes sense as angular momentum. So we're going to try to develop that algebra like that as well here.
You could take it to be an assumption. And as I said, an experiment doesn't tell you the unique way to invent the mathematics. You try to invent the consistent mathematics and see if it coincides with the experiment. And this is a very natural thing to try to invent.
So what are we facing? We're facing a slightly nontrivial problem of figuring out these operators. And they should be Hermitian. So let's try to think of Hermitian 2 by 2 matrices.
So here is a Hermitian 2 by 2 matrix. I can put an arbitrary constant here, because it should be invariant under transposition, which
doesn't change this diagonal value in complex conjugation. So C should be real, D should be real. For the matrix to be Hermitian, 2 by 2 matrix,
I can put an A here. And then this A would have to appear here as well. I can put minus ib, and then I would have plus ib here. So when I transpose and complex conjugate, I get this one.
So this matrix with A, B, C, and D real is Hermitian. Hermiticity is some sort of reality condition.
Now for convenience, I will put a 2C and a 2D here. Doesn't change things too much. Now look at what we're talking about. We're talking about the set of Hermitian matrices.
Funnily, you can think of that again as a vector space. Why a vector space? Well, we'll think about it. In a few seconds, it will become clear. But let me just try to do something here that might help us.
We are trying to identify Sx and Sy from here so that this commutation relations hold. Well, if Sx and Sy have anything to do with the identity matrix, they
would commute with everything and would do nothing for you. So I will remove from these matrices that I'm trying to understand something having to do with the identity. So I'll remove a Hermitian matrix, which is C plus D times the identity, the 2
by 2 identity matrix. This is a Hermitian matrix as well. And I can remove it, and then this matrix is still Hermitian. And this piece that I've removed doesn't change commutators as they appear on the left-hand side.
So if you have an Sx and an Xy here and you're trying to do a computation, it would not contribute. So might as well just get rid of them. So if we remove this, we are left with, you're subtracting C plus D from the diagonal.
So here you'll have C minus D. Here you'll get D minus C, A minus Ib, and A plus Ib. And we should keep searching for Sx and Sy among these matrices.
But then you say, look, I already got Sz, and that was Hermitian. And Sz was Hermitian, and it had a number and the opposite number on the other diagonal entry.
If Sx and Sy have a little bit of Sz, I don't care. I don't want. I want this to be independent matrix. I don't want to confuse the situation. So if this thing has something along Sz, I want it out. So since precisely this number is opposite to this one,
I can add to this matrix some multiple of Sz and kill these things in the diagonal. So add a multiple, add an Sz multiple,
and we finally get this matrix, 0, A minus Ib, A plus Ib, and 0.
So we've made quite some progress. Let's see now what we have. Well, that matrix could be written
as A times 0, 1, 1, 0 plus B times 0 minus i, i, 0.
Which is to say that this Hermitian matrix types a real number, and this Hermitian matrix times a real number. And that makes sense, because if you take a Hermitian matrix and multiply it by a real number, the matrix is still Hermitian. So this is still Hermitian because B is real.
This is still Hermitian because A is real. And if you add Hermitian matrices, it's still Hermitian. So look, in some sense, the set of Hermitian matrices, two by two Hermitian matrices, is a real vector space
with four basis vectors. One basis vector is this. Another basis vector is this. The third basis vector is the Sz part. And the fourth basis vector is the identity that we subtracted.
And this was, I'm listing the other two that we got rid of, because physically we're not that interested, given that we want Sx and Sz. So Sx and Sy. But here it is. These four two by two matrices are
sort of the linearly independent Hermitian matrices. You can think of them as vectors, four basis vectors. You multiply by real numbers and now you add them and you got the most general Hermitian matrix.
So this is part of the subtlety of this whole idea of vector spaces. So matrices can be thought of as vectors sometimes as well. So that's why these matrices are quite famous. But before we just discuss why they are so famous,
let's think of this. We were looking for Sx and Sy. And we actually seem to have two matrices here that could do the job as two independent Hermitian two by two matrices.
But we must add a little extra information. We don't know what the scale is. Should I multiply this by five and call that Sx or this by three? We're missing a little more physics.
What is the physics? The eigenvalues of Sx should also be plus minus h over two. And the eigenvalues of Sy should also be plus minus h over two, just like for Sz. You could have started the whole Stern-Gerlach thing thinking of x.
And you would have obtained plus minus h over two. So that is the physical constraint. These matrices, I have to figure out those numbers. Maybe Sx is this one, Sy is this one. And you can say, oh, you never told us if you're going to get a unique answer here.
And yes, I did tell you. And you're not going to get a unique answer. There's some signs, issues, and some other things. But any answer is perfectly good. So once you get an answer, it's perfectly good. Of course, we're going to get the answer that everybody likes. And the conventions, happily everybody
uses the same convention. Question? So I have a related question because at the beginning we could have chosen the top right entry to be A plus IV and the bottom left to be A minus IV. That would have yielded a different basis matrix. Right, I would have called this plus and minus.
So are we going to show later that this is the correct form? It's not the correct form. It is a correct form. And it's equivalent to any other form you could find. That's what we can show. In fact, I will show there's an obvious ambiguity here.
Well, in fact, maybe I can tell it to you. If you let Sx go to minus Sy and Sy goes to plus Sx, nothing changes in these equations. They become the same equations.
Sx would become minus Sy and this Sx. And this is not changed. But in fact, if you put minus Sy and Sx is the same commutator, then this one will become actually this commutator. And this one will become that. So I could change whatever I get for Sx,
change it for minus Sy, for example, and get the same thing. So there are many changes you can do. The only thing we need one answer that works. And I'm going to write, of course, the one that everybody likes. But don't worry about that. So let's think of eigenvectors and eigenvalues.
Now, I don't know how much you remember that. But we'll just take it at this moment that you do. So 0, 1, 1, 0 has two eigenvalues. A lambda equal 1 with eigenvector 1 over square root of 2, 1, 1.
And a lambda equal minus 1 with eigenvector 1 over square root of 2, 1, minus 1.
The other one is equally easy to do. We will discuss eigenvectors and eigenvalues later.
Minus i, i, 0, 0 has a lambda equal 1 eigenvector with components 1 over square root of 2, 1, and i. Pretty sure it's 1 and i. Yes, and lambda equals minus 1 with components 1
over square root of 2, 1, minus i. Now, I put the 1 over square root of 2 because I want them to be normalized. Remember how you're supposed to normalize these things? You're supposed to take the row vector, complex conjugate,
and multiply. Well, you would get 1 for the length of this, 1 for the length of this. You would get 1 for the length of this. But don't remember, you have to complex conjugate. Otherwise, you'll get 0. Also, you will get 1 for the length of this.
So these are our eigenvalues. So actually, with eigenvalues lambda equal 1 and minus 1 for this 2, we're in pretty good shape. We could try Sx to be h bar over 2 times 0, 1, 1, 0,
and Sy to be h bar over 2, 0, minus i, i, 0. This would have the right eigenvalues because if you multiply a matrix by a number,
the eigenvalue gets multiplied by this number. So the plus minus 1's become plus minus h bar over 2. Well, what are we supposed to check? If this is to work, we're supposed to check these commutators. So let's do 1 at least, Sx commutator with Sy.
So what do we get? h bar over 2, h bar over 2, two of them. Then the first matrix, 0, 1, 1, 0, times 0, minus i, i, 0,
minus 0, minus i, i, 0, 0, 1, 1, 0, which is h bar over 2 times h bar over 2 times of, you get i, 0, 0, minus i,
minus, minus i, 0, 0, i.
And we're almost there. What do we have? Well, we have h bar over 2, h bar over 2.
And we got 2i and minus 2i. So this is h bar over 2 times h bar over 2 times 2i times 1, minus 1.
And this whole thing is ih bar. And the other part is h bar over 2, 1, minus 1, which is ih bar Sz hat. Good, it works.
The only thing that could have gone wrong, you could have identified one with a minus or other like that. It would have been equally good once you have these operators were fine. So one has to check that the other ones work, and they do.
I will leave them for you to check. And therefore, we've got the three matrices. It's a very important result, the Sx, Sy, and Sz. So I will not rewrite them, but they
should be boxed nicely, the three of them together with that one there, top of the blackboard. And of course, by construction, they're Hermitian. They're famous enough that people have defined
the following object. Si is defined to be h bar over 2 sigma i, the Pauli matrix sigmas. And these are Pauli matrices. Sigma 1 is 0, 1, 1, 0. Sigma 2 is 0, minus i, i, 0.
And sigma 3 is equal to 1, minus 1, 0, 0. OK, so in principle, yes, question? Is it at all significant that the Pauli matrix
is all squared to be a matrix? Yes, it is significant. We'll use it, although at this moment, it's not urgent for us. We'll have no application of that property
for a little while, but it will help us do a lot of the algebra of the Pauli matrices. Sorry?
That's right. I think so. Our eigenvalues, yeah, it's true.
The fact that the eigenvalues are plus minus 1 will imply that these matrices square to themselves. So it's incorporated in our analysis. The thing that I was saying is that you don't sort of
need it in the expression of the commutators. So in the commutators, it didn't play a role to begin with. Put it as an extra condition. Now, what is the next thing we really want to understand? It's in terms of plane states, we now
have the answer for most of the experiments we could do. So in particular, remember that we said that we would have Sx, for example, having states x plus minus, which are h bar over 2 plus minus x comma plus minus.
So the states along the x direction, referred like that, would be these eigenstates of the Sx operator. But we've calculated the eigenstates of the Sx operator.
They're here. The Sx operator is h bar over 2 times this matrix. And we have those things. So the plus eigenvalue and the minus eigenvalue will just show up here. So let me write them and explain in plain language
what the states are. So the eigenstate with lambda equal 1, that would correspond to h bar over 2. So the x plus corresponds to this vector.
So what is that state? It's that vector, which, if you want more explicitly, it's the z plus plus z minus.
This is the state 1 over square root of 2, 1, 1. The x minus is z plus minus z minus.
As you see on that blackboard, it's 1 minus 1. So here it is. The states that you were looking for that are aligned along x, plus x or minus x, are not new states that you have to add to the state space.
They are linear combinations of the states you got. We can invert this formula and write, for example, that z plus is 1 over square root of 2 x plus plus 1
over square root of 2 x minus. And z minus is 1 over square root of 2 x plus minus 1 over square root of, oops, minus the square root of 2
is already out, I'm sorry, minus x minus. So actually, this answers the question that you had. For example, you put a z plus state,
and you put an x filter. What amplitude do you have to find a state in the x plus, given that you start with a state on the z plus?
Well, you put an x plus from here. You get 1 from this and 0 from this 1, because the states are always orthogonal. These states are orthogonal. You should check that. And therefore, this is 1 over square root of 2.
If you ask for x minus with respect to z plus, that's also 1 over square root of 2. And these are the amplitude for this state to be found in this. For this state to be found in them, they're equal, the probabilities are 1 half. And that's good.
Our whole theory of angular momentum has given us something that is perfectly consistent with the Stern-Gerlach experiment and gives you these probabilities. You can construct in the same way the y states.
So the y states are the eigenstates of that second matrix Sy that we wrote on the left. So this matrix is Sy.
So it's eigenstates. I'm sorry. It's there. Sy is there. So the eigenstates are those. So immediately, you translate that to say that Sy has eigenstates y plus minus.
Eigenvalues are plus minus h bar over 2 y plus minus. And y plus is equal 1 over square root of 2 z plus. And look at the first eigenvector.
Plus i z minus. And in fact, I can put one formula for both. Here they are. So it's kind of neat that the x1's
were found by linear combinations and they're orthogonal. Now, if you didn't have complex numbers, you could not form another linear combination that is orthogonal. But thanks to these complex numbers, you can put an i there. There's no i in the x1's.
And these states are orthogonal, something that you should check. So again, you can invert and find the z states in terms of y's. And you would conclude that the amplitudes are really the same up to signs or maybe complex numbers. But the probabilities are identical.
So we've gotten a long way. We basically have a theory that seems to describe the whole results of the Stern-Gerlach experiment. But now your theory can do more for you. In the last few minutes, we're going to calculate the states that are along arbitrary directions.
So here I produced a state that is along the x direction plus and along the x direction minus. What I would like to construct to finish this story
is a state that is along some arbitrary direction. So the state that points along some unit vector n. So here is space. And here is a unit vector n with components nx, ny, and nz.
Or you can write the vector n as nx ex plus ny ey plus nz ez.
And I would like to understand how I can construct in general a spin state that could be said to be in the n direction. We have the ones along the z, x, and y. But let's try to get something more general, the most general one.
So for this, we think of the triplet of operators s, which would be sx, sy, and sz. Now you can, if you wish, write this
as sx hat ex vector plus sy hat ey vector plus sz hat ez vector. But this object, if you write it like that, is really a strange object.
Think of it. It's matrices or operators multiplied by unit vectors. These vectors have nothing to do with the space in which the matrices act. The matrices act on an abstract two-dimensional vector space,
while these vectors are sorted for accounting purposes. That's why we sometimes don't write them and say we have a triplet. So this product means almost nothing. They're just sitting together. You could put the e to the left of the x or to the right. It's a vector. You're not supposed to put the vector inside the matrix
either. They don't talk to each other. It's an accounting procedure that is useful sometimes. We will use it to derive identities soon. But it's an accounting procedure. So here's what I want to define. So this is a crazy thing, some sort
of vector-valued operator or something like that. But what we really need is what we'll call s hat n, which will be defined as n dot s, where we take naively what a dot product is supposed to mean.
This component times this component, which happens to be an operator. This times this, this times that. So nx sx plus ny sy plus nz sz. And this thing is something very intuitive.
It is just an operator. It doesn't have anymore a vector with it. So it's a single operator. If your vector points in the z direction, nx and ny is 0, you have Sz, because it's a unit vector.
If the vector points in the x direction, you get Sx. If the vector points in this y direction, you get Sy. In general, this we call the spin operator in the direction of the vector n.
Spin operator in the direction of n. OK, so what about that spin operator? Well, it had eigenvalues plus minus h bar over 2 along z, x,
and y. Probably it does still have those eigenvalues. But we have to make this a little clearer. So for that, we'll take nx, ny, and nz
to be the polar coordinate thing. So this vector is going to have a theta here and the azimuthal angle phi over here. So nz is cosine theta. nx and ny have sine theta.
And nx has cos phi, and this one has sine phi. So what is the operator Sn vector hat? Well, it's nx times Sx.
So I'll put an h bar over 2 in front. So we'll have nx sigma x, or sigma 1, plus ny sigma 2, plus nz sigma 3.
Remember, the spins operators are proportional, h bar over 2 times the sigmas. So sigma 1, sigma 2, sigma 3. And look what we get, h bar over 2. Sigma 1 has an nx here, nx.
Sigma 2 has minus i ny plus i ny. And sigma 3 would have an nz minus nz. So this is h bar over 2.
nz is cosine theta. nx minus i ny, oh, it's a pretty awful thing, but it's very simple. nx minus i ny is sine theta times e to the minus i phi.
Here it would be sine theta e to the i phi. And here we'll have minus cosine theta. So this is this whole matrix, Sn hat like that.
Well, in the last couple of minutes, let's calculate the eigenvectors and eigenvalues. So what do we get? Well, for the eigenvalues, remember what is the computation of an eigenvalue of a matrix.
It's an eigenvalue of a matrix A. You write by solving determinant of A minus lambda 1 equals 0. So for any matrix A, so if we want
to find the eigenvalues of this matrix, we would have to write eigenvalues of Sn hat. We would have to write the determinant of this minus lambda i. So the determinant of h bar over 2 cosine theta minus lambda.
Minus h bar over 2 cosine theta minus lambda. And here, sine theta e to the minus i phi.
Sine theta e to the i phi. The determinant of this being 0. It's not as bad as it looks. It's actually pretty simple. These are A plus B, A minus B. Here, the faces cancel out.
The algebra, you can read it in the notes. But you do get lambda equal plus minus h bar over 2. Now, that is fine. And we now want the eigenvectors.
Those are more non-trivial. So we need a little more work. So what are you supposed to do to find an eigenvector? You're supposed to take this A minus lambda i acting on a vector and put it equal to 0.
And that's the eigenvector. So for this case, we're going to try to find the eigenvector N plus. So this is the one that has S N on this state.
Well, I'll write it here. Plus minus h over 2 N plus minus here. So let's try to find this one that corresponds to the eigenvalue equals 2 plus h bar over 2.
Now, this state is C1 times z plus plus C2 times z minus. These are our basis states. So it's a little combination.
Or it's C1, C2. Think of it as a matrix. So we want the eigenvectors for that. So what do we have? Well, we would have S N hat minus h bar over 2 times 1
on this C1, C2 equals 0. The eigenvector equation is that this operator minus the eigenvalue must give you that.
So the h bars over 2 happily go out. And you don't really need to worry about them anymore. And you get here cosine theta minus 1 sine theta e to the minus i phi sine theta e to the i phi
and minus cosine theta minus 1 C1, C2 equals 0. All right. So you have two equations.
And both relate C1 and C2 happily. And the reason this works is because with this eigenvalue that we've used that appears here, these two equations are the same. So you can take either one.
And they must imply the same relation between C1 and C2, something you can check. So let me write one of them. C2 is equal to e to the i phi 1 minus cosine theta over sine theta C1.
It's from the first line.
So you have to remember, in order to simplify these things, your half angle identities. Sorry. Sorry. 1 minus cosine theta is 2 sine squared theta over 2.
And sine theta is 2 sine theta over 2 cosine theta over 2. So this becomes e to the i phi sine theta over 2 over cosine theta over 2 C1.
Now, we want these things to be well normalized. So we want C1 squared plus C2 squared equal to 1. So you know what C2 is. So this gives you C1 squared times 1 plus, and C2,
you use this. When you square, the phase goes away. Sine squared theta over 2 cosine squared theta over 2 must be equal to 1. Well, the numerator is 1, so you learn that C1 squared is equal to cos squared theta over 2.
Now, you have to take the square root, and you could put an i or a phase or something. But look, whatever phase you choose, you could choose C1 to be cosine theta over 2
and say, I'm done. I want this one. Somebody would say, no, let's put the phase e to the i pi over 5. So it doesn't look good. What for? Even worse, this phase will show up in C2
because C2 is proportional to C1. So I can get rid of it. I only should put it if I really need it. And I don't think I need it, so I won't put it. And you can always change your mind later. Nobody's going to take your word for this. So in this case, C2 would be sine theta over 2 e
to the i pi. It's nice because it cancels. And therefore, we got the state. n plus is supposed to be cosine theta over 2 z plus and plus
sine theta over 2 e to the i pi z minus. This is a great result. It gives the arbitrarily located spin state that points in the n direction
as a linear superposition of your two basis states and answers conclusively the question that any spin state in your system can be represented in this two-dimensional vector space. Now moreover, if I would take theta equals 0,
if I take theta equals 0, I have the z-axis. And it's independent of the angle phi. The phi angle becomes singular at the North Pole. But that's all right. When theta is equal to 0, this term is 0 anyway. And therefore, this goes. And when theta is equal to 0, you recover the plus state.
Now you can calculate the minus state. And if you follow exactly the same economical procedure, you will get the following answer. And I think unless you've done a lot of eigenvalue
calculation, this is a calculation you should just redo it. So the thing that you get without thinking much is that n minus is equal to sine theta over 2 plus minus cosine theta over 2 e to the i phi minus.
At least some way of solving this equation gives you that. And you could say this is natural, and this is fine. But it's not so nice, actually. Take theta equal to pi.
Then theta equal to pi. So again, you take theta equal to 0.
Theta equal to 0, this is supposed to be the minus state along the direction. So this is supposed to give you the minus state, because the vector n is along up in the z direction. But you're looking at the minus component. So theta equals 0.
Sure, there's no plus. But theta equals 0, you get the minus state. And this is 1, and phi is ill-defined. But it's not so nice, therefore. So at this moment, it's convenient to multiply this state by e to the minus i phi times minus 1.
Just multiply it by that. So that n minus is equal to minus sine theta over 2 e to the minus i phi plus cosine theta over 2 minus.
And that's a nice definition of the state. When theta is equal to 0, you're fine. And it's more naturally equivalent to what you know. Theta equal to 0, you get the minus state, or z minus.
I didn't put the z's here for laziness. And for theta equal to 0, anyway, the phase phi doesn't matter. So it's a little nicer. You could work with this one, but might as well leave it like that. So we have our general states.
We've done everything here that required that some linear algebra, without doing our review of linear algebra. But that's what we'll start to do next time.