Richard Seto
Lecture 12
Updated for 2005
Date@D
Lecture 12
Changing basis (or "representation")
We have decided to use the Sz basis, i.e. †+\ and †-\. Suppose we want to switch to the Sx basis. Is the a way to do this?
Before we answer this, lets see if there is a way to go from one set of eigenkets say †+\ and †-\, to another set, perhaps
†Sx +\ and †Sx -\. Lets think of this more generally as A=S z with eigenkets †ai \. Is there a way to find a different set of
` `
eigenkets of another operator B=S x with eigenkets †bi \? In general we can let A and B be any two operators which
` ` ` `
represent observables so that the kets †ai \ and †bi \are complete, orthogonal eigenkets. Let us suppose that we have an
`
operator U which does this.
`
The operator U will be a special kind of transformation matrix called a unitary operator.
` `† ` ` ` `† ` ` -1 ` †
A unitary operator U is one in which U U =1 and U U =1 This also means U =U (2)
Now we want an operator U such that †bi \=Ù†ai \. By construction this will be ⁄k †bk \ Xak »
`
(3)
we can also show that this is unitary: U U =⁄ j †a j \ Xb j »⁄k †bk \ Xak »=⁄ j ⁄k †a j \Xb j †bk \ Xak »=⁄ j †a j \Xa j »=1̀
`† `
`
Now we can think of U in a matrix notation. The matrix elements are (see lecture 7)
where we have used (3). Now lets see how we can use this.
We have typically written things in the Sz basis, but suppose we want to switch to a different basis, say Sx . How do we do
this?
First lets start out by figuring out how to change the basis of kets. In the Sz basis we write †a\=†+\X+†a\+ †-\X-†a\.
So that we can write this symbolically as †a\=⁄i †ai \ ci = ⁄i †ai \ Xai †a\ where the †ai \ are eigenkets of A and A=S z . We see
` ` `
that the coefficients ci = Xai †a\. Now lets let B=S x and †bi \ are the eigenkets. In general we can let A and B be any two
` ` ` `
operators which represent observables so that the kets †ai \ and †bi \are complete, orthogonal eigenkets. Remember
‚ †ai \ Yai ° = 1 and ⁄i †bi \ Xbi †=1̀ Then we can write
`
†a\=⁄i †ai \ Xai †a\=⁄ij †b j \ Xb j †ai \ Xai †a\=⁄ j †b j \ ⁄i Xb j †ai \ Xai †a\ So we can now write †a\ = ⁄ j †b j \ c' j where the new
i
coefficients Xb j †a\=c' j =⁄i Xb j †ai \ Xai †a\. Thinking of this in matrix notation is particularly nice since originally in the A
`
i Xa1 †a\ yz
basis we write a column vector jj z
k Xa2 †a\ {
i Xb1 †a\ yz ij Xb1 †a1 \ Xb1 †a2 \ yzij Xa1 †a\ yz
We can transform to the new basis as follows: jj z=j zj z.
k Xb2 †a\ { k Xb2 †a1 \ Xb2 †a2 \ {k Xa2 †a\ {
(5)
Notice how this matrix multiplication does exactly the same thing as the sum. Now notice that this matrix is just the
`
transpose-complex conjugate of U so we can write
`†
(new basis)=U (old basis) (7)
Example:
è!!!
ÅÅÅÅ!ÅÅ (†Sx +\ +†Sx -\ ) è!!!
ÅÅÅÅ!ÅÅ (†Sx +\ -†Sx -\ )
1 1
(b)
è!!!
ÅÅÅÅ!ÅÅ (†S y +\ +†S y -\ ) è!!!!ÅÅÅÅÅ (†S y +\ -†S y -\ ) (c)
1 1
1) find the representation of †±\ in the †Sx ±\ basis. We know what the answer should be just by looking at (b).
1 i 1y 1 i 1 y
It should be †+\ = ÅÅÅÅ è!!! ÅÅÅÅ!ÅÅ jj zz and †-\ = ÅÅÅÅ
ÅÅÅÅ!ÅÅ (†Sx +\ +†Sx -\ ) U ÅÅÅÅ
è!!! è!!! ÅÅÅÅ!ÅÅ jj zz in the †Sx ±\ basis.
ÅÅÅÅ!ÅÅ (†Sx +\ -†Sx -\ ) U ÅÅÅÅ
è!!!
2 k1{ 2 k -1 {
1 1
Its very important that we recall we are in the †Sx ±\ basis! Lets now use what we learned to see if we get the
2 2
` i Xold1 †new1 \ Xold1 †new2 \ yz ij X+†Sx +\ X+†Sx -\ yz 1 i X+†H†+\ + †-\L X+†H†+\ - †-\L y 1 i 1 1 y
same answer.
ÅÅÅÅ!ÅÅ j
2 1 2 x x
è!!!
2 k 1 -1 { k0{ k1{
U = ÅÅÅÅ
1 i 1 1 yi 1 y 1 i 1 y 1 i 1 1 yi 0 y 1 i 1 y
ÅÅÅÅ!ÅÅ jj
†+\U ÅÅÅÅ
è!!!
zzjj zz= ÅÅÅÅ
ÅÅÅÅ!ÅÅ jj zz
è!!! ÅÅÅÅ!ÅÅ jj
†-\U ÅÅÅÅ
è!!!
zzjj zz= ÅÅÅÅ
ÅÅÅÅ!ÅÅ jj zz in the new basis as we expected.
è!!!
2 k 1 -1 {k 0 { 2 k1{ 2 k 1 -1 {k 1 { 2 k -1 {
`
How does this change the matrix elements of operators? Lets find out for an operator X , Remembering that
†bi \=Ù†ai \ and hence Xbi »=Xai »Ù
†
Ybk » X » bl ]=Ybk … 1 X 1 … bl ]=Ybk … ‚ †ai \ Yai °X̀ » ⁄ j †a j \ Xa j †bl \=⁄i ⁄ j Xak »Ù » ai \ Yai °X̀ » ⁄ j †a j \ Xa j »Ù†al \=
` ` ` ` †
=Xak »Ù X U †al \ or in somewhat symbolic notation where we mean that X ' is just the matrix for X in a new basis or
i
† ` ` ` `
representation
` `† ` `
X '=U X U - This is known as a similarity transformation in matrix algebra.
`† ` `† ` `
So to summarize (ket in new basis)=U (ket old basis) AND X '=U X U (8)
i 0 1 yz
2) Now Sx in the †±\ basis is ÅÅÅÅÑ2 jj
Example:
z. Lets write it in the †Sx ; ≤\ basis. We again know what the answer has to
`
k1 0 {
i 1 0 yz
be - it has to be ÅÅÅÅÑ2 jj z. Lets use our methods and see if that is what we get. We use (8)
k 0 -1 {
1 i 1 1 y i 0 1 y i 1 1 y Ñ 1 i 1 1 y i 1 -1 y Ñ 1 i 2 0 y Ñ i 1 0 y
ÅÅÅÅ!ÅÅ jj zz jj zz jj zz = ÅÅÅÅ ÅÅÅÅ jj zz jj zz = ÅÅÅÅ ÅÅÅÅ jj zz= ÅÅÅÅ jj zz as we
`† ` `
è!!!
ÅÅÅÅ!ÅÅ ÅÅÅÅ
è!!!
2 k 1 -1 { k 1 0 { k 1 -1 { 2 2 k 1 -1 { k 1 1 { 2 2 k 0 -2 { 2 k 0 -1 {
U Sx U = ÅÅÅÅÑ2 ÅÅÅÅ
1
2
expected.
TRACE
Proof: ‚ Yai » X » ai ]=‚ Yai … ‚ †bk \ Ybk °X̀ » ⁄l †bl \ Xbl †ai \=‚ ‚ ‚ Xai » bk \ Ybk °X̀ » †bl \Xbl †ai \=
`
=‚ ‚ ‚ Ybl †ai \ Xai » bk \ Ybk °X̀ » bl ]=‚ ‚ Ybl °1̀ » bk ] Ybk °X̀ » bl ]=‚ ‚ dkl Ybk °X̀ » bl ]=⁄k Ybk °X̀ » bk ] QED
i i k k l i
k l i k l k l
A few other things I will ask you to prove in the problem set
Tr(X Y L=Tr(Ỳ X )
` ` `
Tr(U X U L = TrHX )
`† ` ` `
TrH†ai \ Xa j §L = dij
TrH†bi \ Xai §L =Xai †bi \
(10)
i 0 1 yz i 1 0 yz
3) Now in example 2 we had two representation of Sx as ÅÅÅÅÑ2 jj z in the ususal †±\ and as ÅÅÅÅÑ2 jj
Example:
z in the
`
k { k 0 -1 {
†Sx ; ≤\. The trace is the same (i.e. zero) in either basis.
1 0
Example:
4) now lets try something harder. Let's write †Sx ; ≤\ in the †S y ; ≤\ basis. So we have
1 i 1y 1 i 1 y
In the original †±\ basis » Sx +\ U ÅÅÅÅ ÅÅÅÅ!ÅÅ jj zz and » Sx -\ U ÅÅÅÅ
è!!! ÅÅÅÅ!ÅÅ jj zz
è!!!
2 k1{ 2 k -1 {
1 i 1 -i yi 1 y 1 i 1 - i y ij 1 -i yzij 1 yz 1 ij 1 + i yz
In the new basis » Sx +\U ÅÅÅÅ è!!! ÅÅÅÅ!ÅÅ jj
ÅÅÅÅ!ÅÅ ÅÅÅÅ
è!!! zzjj zz= ÅÅÅÅ jj zz †-\U ÅÅÅÅ è!!!
ÅÅÅÅ!ÅÅ ÅÅÅÅ
è!!!
ÅÅÅÅ!ÅÅ j zj z= ÅÅÅÅ j z
2 k 1 i {k 1 { 2 k 1 + i { k 1 i {k -1 { 2 k 1 - i {
1 1 1
2 2 2
i 0 1 yz
2) Now Sx in the †±\ basis is ÅÅÅÅÑ2 jj z. Lets write it in the †S y ; ≤\ basis.
`
k1 0 {
1 i 1 -i y i 0 1 y i 1 1 y Ñ 1 i 1 -i y i i -i y Ñ 1 i 0 -2 i y Ñ ij 0 -i yz
ÅÅÅÅ!ÅÅ jj zz jj zz jj zz = ÅÅÅÅ ÅÅÅÅ jj zz jj zz = ÅÅÅÅ ÅÅÅÅ jj zz= ÅÅÅÅ j z is
`† ` `
è!!!
ÅÅÅÅ!ÅÅ ÅÅÅÅ
è!!!
2 k 1 i { k 1 0 { k i -i { 2 2 k 1 i { k1 1 { 2 2 k2i 0 { 2 ki 0 {
U Sx U = ÅÅÅÅÑ2 ÅÅÅÅ
1
Diagonalization:
Suppose we know the matrix elements Yai °B̀ †a j \ where the †a j \ are not the eigenkets of B. How do we find the basis in
`
which a matrix is diagonal? i.e. we want to find B in the †b j \ basis. OK - so let's first lets look at an operator B in the
` `
Ybi » B » b j ] = b j Xbi » b j \ = b j dij So in this basis B is diagonal. If we find the unitary operator where
` `
†bi \ = U †ai \ then we know that U BU is diagonal. So we have to find U such that †bi \ = U †ai \ and the †bi \ are
` `†` ` ` `
`
eigenvectors of B.
So first we want to find the eigenvectors and eigenvalues of an operator B where we know the matrix elements Yai °B̀ †a j \
`
and where the †a j \ are not the eigenkets of B. This means we want to find the b j 's such that
`
writing this in matrix notation for a 2x2 for example we have two eqn, one for each eigenvalue b j = b1 and b2
det(B-l1)= 0 where the values of l will be the eigenvalues b j . Once we solve for the b j we can find the eigenvectors †b j \
` `
i Xa1 » b j \ yz
which in matrix notation is jj z . Now define xij = Xai » b j \ . But from (4) xij is just the expression for the unitary
k Xa2 » b j \ {
operator we want since Xai »Ù†a j \= Xai »b j \. So all we have to do is to put the eigenvectors †b j \ in the †ai \ basis, side by
` ` `
i 0 1 zy ` ` i Xa1 » b j \ zy i0y
Ñjj z and the procedure will fail since (B-l1)jj z= 0 just give 0=0. (This is since one of the eigenkets is jj zz).
side and we get U . Note that for this procedure to work it is important that B be hermitian. e.g. S + in the Sz basis is
k0 0 { k Xa2 » b j \ { k0{
Example:
`
5) now lets start with S x in the Sz representation and change it to the Sx representation where it should be diagonal.
First we have that the †ai \=†±\ and the †bi \ = †Sx ≤\ = ÅÅÅÅ è!!!
ÅÅÅÅ!ÅÅ (†+\±†-\)
` ij Xa1 †b1 \ Xa1 †b2 \ yz ij X+†Sx ; +\ X+†Sx ; -\ yz 1 ij +1 +1 yz
1
` † 1 ij +1 +1 yz
U =j z=j ÅÅÅÅ!ÅÅ j
z= ÅÅÅÅ z ÅÅÅÅ!ÅÅ j z
2
` † ` 1 ji +1 +1 yz 1 ji +1 +1 yz i +2 0 zy `
check unitarity
ÅÅÅÅ!ÅÅ j
è!!! ÅÅÅÅ!ÅÅ j
z ÅÅÅÅ
è!!! z = ÅÅÅÅ12 jj z= 1 it is
2 k +1 -1 { 2 k +1 -1 { k 0 +2 {
U U = ÅÅÅÅ
1 i +1 +1 y i 1 y 1 ij 1 yz 1 i +1 +1 y i 0 y 1 i 1 y
check that
†b j \ = U †ai \ ÅÅÅÅ!ÅÅ jj zz jj zz= ÅÅÅÅ j zU†Sx ; +\ and ÅÅÅÅ ÅÅÅÅ!ÅÅ jj zz jj zz= ÅÅÅÅ
ÅÅÅÅ!ÅÅ jj zzU†Sx ; -\ good
`
è!!! è!!!
ÅÅÅÅ!ÅÅ è!!! è!!!
2 k +1 -1 { k 0 { k { k +1 -1 { k 1 { 2 k -1 {
ÅÅÅÅ
1
i 1 0 yz
2 2
ÅÅÅÅÑ2 jj z. That is we expect that the †b j \ are eigenkets of U S x U with the same eigenvalues as the S x eigenvalues.
`†` ` `
k 0 -1 {
` † ` ` 1 ij +1 +1 yz Ñ ij 0 1 yz 1 ij +1 +1 yz 1 Ñ ij +1 +1 yz ij +1 -1 yz 1 Ñ ij 2 0 yz Ñ ij 1 0 yz
ÅÅÅÅ!ÅÅ j z ÅÅÅÅ j ÅÅÅÅ!ÅÅ j
z ÅÅÅÅ z= ÅÅÅÅ ÅÅÅÅ j zj z= ÅÅÅÅ ÅÅÅÅ j z= ÅÅÅÅ j z so it
`
è!!! è!!!
2 k +1 -1 { 2 k 1 0 { 2 k +1 -1 { 2 2 k +1 -1 { k +1 +1 { 2 2 k 0 -2 { 2 k 0 -1 {
S x '=U S x U = ÅÅÅÅ
i 1 0 yz ij 0 1 yz
works!
` i 1 -1 zy
Diagonalize X = jj
Example:
z
k -1 1 {
So we need to find the eigenvalues and eigenkets of this matrix. The eigenkets will
1 i 1y
ÅÅÅÅ!ÅÅ jj zz
è!!!
2 k1{
for b=0 a-b=0 a=-b and using a2 + b2 = 2 we get that the eigenket is ÅÅÅÅ
1 j i 1y
ÅÅÅÅ!ÅÅ j zz
è!!!
2 k -1 {
for b=2 -a-b=0 and a=-b and using a2 + b2 = 2 we get that the eigenket is ÅÅÅÅ
So old basis is †±\ the new basis is ÅÅÅÅ ÅÅÅÅ!ÅÅ H†+\+†-\) and ÅÅÅÅ
è!!!
1
ÅÅÅÅ!ÅÅ H†+\-†-\)
è!!!
1
` 1 ij yz ` † 1 ij yz
ÅÅÅÅ!ÅÅ j z and U = ÅÅÅÅÅÅÅÅ!ÅÅ j z
2 2
è!!! è!!!
2 k 1 -1 { 2 k 1 -1 {
1 1 1 1
U = ÅÅÅÅ
` -1 ` ` -1 `
Note first that the definition of U is the inverse of U so that U U =1
We can by definition write B †bl ] = bl † bl \ and comparing this with U A U †bl ] = B †bl ] = al † bl \ we recognize
` ` ` ` -1 `
` ` ` ` -1
that B and U A U are simultaneously diagonalizable with the same eigenvalues. Often in the cases of physical interest,
they are the same operator.
` 1 ij +1 +1 yz ` -1 ` † 1 ij +1 +1 yz
In our example †bi \ = †Sx ≤\ B = S x † ai \=†±\ A = S z and U = ÅÅÅÅ ÅÅÅÅ!ÅÅ j z U = U = ÅÅÅÅ ÅÅÅÅ!ÅÅ j z
` ` ` `
è!!! è!!!
2 k +1 -1 { 2 k +1 -1 {
1 j i +1 +1 zy Ñ ji 1 0 zy 1 ji +1 +1 zy 1 Ñ ji +1 +1 zy ji +1 +1 zy 1 Ñ ji 0 2 zy Ñ ji 0 1 zy
ÅÅÅÅ!ÅÅ j z ÅÅÅÅ j ÅÅÅÅ!ÅÅ j
z ÅÅÅÅ z= ÅÅÅÅ ÅÅÅÅ j zj z= ÅÅÅÅ ÅÅÅÅ j z= ÅÅÅÅ j z
` ` ` -1
è!!! è!!!
2 k +1 -1 { 2 k 0 -1 { 2 k +1 -1 { 2 2 k +1 -1 { k -1 +1 { 2 2 k 2 0 { 2 k 1 0 {
U S z U = ÅÅÅÅ
` ` i 0 +1 yz i1y i 1 y
Now B = S x Ujj z and has eigenvalues ± ÅÅÅÅÑ2 with eigenvectors jj zz and jj zz Lets check and see if these are
k { k { k -1 {
` ` ` -1 Ñ ij 0 1 yz
+1 0 1
eigenvectors of U S z U = ÅÅÅÅ2 j z This is just S x again, so its obvious that they are. As stated, the two operators
`
k1 0 {
` ` ` ` -1
B and U A U are often the same operator in cases of interest.
` 1 ij +1 +1 yz
ÅÅÅÅ!ÅÅ j z and
` ` ` ` ` -1
è!!!
2 k +1 -1 {
So what this says in this case is that S z and S x =U S z U are related by a unitary transformation U = ÅÅÅÅ
≤Ñ ` `
have the same eigenvalues ÅÅÅÅ ÅÅÅÅ . Now what is interesting is that we know what sort of transformation changes S z to S x - its
` 2
a rotation. But is U really a rotation? Yes it is. It tells us how the 2 dimensional space of spin transforms under rotation in
3-space.
Note of interest:
` `
The transformation that changes S z to S x is
` ` ` ` -1 `
S i zU S i U is the same as Rij S j where Rij is just a rotation. This can be done because the U's are just 2x2 matrices. But
actually there are 2 U's that will do it. U, and -U this is just the double covering aspect of SU(2) to SO(3) which then
means that fermions have a funny thing that a rotation in 2p does not bring you back to the same thing but you have to go
through 4p.