* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download Ch 6
Moment of inertia wikipedia , lookup
Field (physics) wikipedia , lookup
Time in physics wikipedia , lookup
Special relativity wikipedia , lookup
Vector space wikipedia , lookup
Minkowski space wikipedia , lookup
Euclidean vector wikipedia , lookup
Nordström's theory of gravitation wikipedia , lookup
Metric tensor wikipedia , lookup
Symmetry in quantum mechanics wikipedia , lookup
Vector Spaces in Physics
8/6/2015
Chapter 6. Rotations and Tensors
There is a special kind of linear transformation which is used to transforms coordinates
from one set of axes to another set of axes (with the same origin). Such a transformation
is called a rotation. Rotations have great practical importance, in applications ranging
from tracking spacecraft to graphical display software. They also play a central role in
theoretical physics
The linear transformation corresponding to a rotation can be described by a matrix. We
will describe the properties of rotation matrices, and discuss some special ones.
A. Rotation of Axes.
In chapter 1 we considered a rotation
of a Cartesian coordinate system
about the z-axis by an angle . A
more general case is shown in figure
6-1, where the axes are transformed
by rotation into an arbitrary second
set of axes. (We take both sets or
axes to be right handed.
Transformation to a left-handed
coordinate system is to be avoided,
unless you really know what you are
doing.) A typical vector A is also
shown.
z
z'
A
y'
eˆ3 '
ê1
ê3
eˆ2 '
y
eˆ1 ' ê2
x'
x
Figure 6-1. Two sets of Cartesian coordinate axes.
The (x',y',z') system is obtained from the (x,y,z)
system by a rotation of the coordinate system.
Figure 6-2 (a) shows the components
of A in the (eˆ1 , eˆ2 , eˆ3 ) system, and
figure 6-2 (b) shows its coordinates in the (eˆ1 , ' eˆ2 , ' eˆ3 ' ) system. We see that we can write
A , using components in the unprimed system, as
A eˆi Ai ;
(6-1)
or, using components in the primed frame, as
A eˆi ' Ai '
(6-2)
The standard definitions of components are clear from (6-1) and (6-2):
Ai eˆi A,
Ai ' eˆi ' A.
We can, however, relate the primed components to the unprimed, as follows:
6-1
(6-3)
Vector Spaces in Physics
8/6/2015
A ' eˆ j ' A
j
eˆ j ' ( Al eˆl )
(6-4)
eˆ j ' eˆl Al
C jl Al
(a)
z
A
z
(b)
eˆ3 A3
A
z'
eˆ3 A3
y'
eˆ3 '
y
eˆ1 A1
ê3 eˆ '
2
ê1
eˆ1 ' ê2
y
eˆ2 A2
eˆ1 A1
eˆ2 A2
x'
x
x
Figure 6-2. The vector A expressed in terms of its (eˆ1 , eˆ2 , eˆ3 ) components (a), and in terms of its
(eˆ1 , ' eˆ2 , ' eˆ3 ' ) components (b).
That is to say that the relationship between primed and unprimed components can be
expressed as a matrix equation,
(6-5)
A' C A ,
where the transformation matrix C is given by
C C jl (eˆ j ' eˆl ) .
(6-6)
There is an interpretation of the coefficients Cjl in terms of angles. Since all of the eˆj
and eˆl are unit vectors, the dot product of any two is equal to the cosine of the angle
between them:
(6-7)
eˆm ' eˆn cosm'n .
Example: Consider a rotation by an angle in the counter-clockwise sense about the z
axis, as described in Chapter 1. Since the z axis is the same in the primed and unprimed
systems, certain of the coefficients C jl eˆ j ' eˆl are especially simple:
C 33 1,
C 31 C 32 C13 C 23 0
The other coefficients can be read off from figure (2-4):
6-2
;
(6-8)
Vector Spaces in Physics
8/6/2015
C11 î 'î cos ,
C 22 ĵ' ĵ cos ,
C12 î ' ĵ sin ,
;
(6-9)
C 21 ĵ'î - sin .
and so
cos sin 0
(6-10)
C sin cos 0 Rz ( ) .
0
0
1
The matrices corresponding to rotations about the x and y axes can be calculated in a
similar way; the results are summarized below.
0
0
1
Rx ( ) 0 cos sin
0 sin cos
cos
Ry ( ) 0
sin
cos
Rz ( ) sin
0
0 sin
1
0 .
0 cos
sin
cos
0
(6-11)
0
0
1
B. Some properties of rotation matrices.
Orthogonality. Consider the definition of the transformation matrix C :
C C jl (eˆj eˆl ) .
(6-6)
If we form the product of C and its transpose, we get
CC T CimC T ml
il
CimClm
eˆi ' eˆmeˆl ' eˆm
.
(6-12)
eˆi ' eˆmeˆm eˆl '
But, the vector eˆl , like any vector, can be written in terms of its components in the unprimed frame,
(6-13)
eˆl eˆm (eˆm eˆl ) .
In this sum, (eˆm eˆl ) is the m-th component of the vector eˆl in the unprimed frame. So,
equation (6-12) becomes
6-3
Vector Spaces in Physics
8/6/2015
CC il eˆi ' eˆl ' .
T
(6-14)
il
or
CC T I .
The equivalent statement in index notation is
Cij Clj il .
(6-15)
(6-15a)
A similar calculation for the product C T C leads to
CT C I
.
CliClj il
(6-15b)
This leads us to define orthogonality as follows:
For an orthogonal matrix, the
transpose matrix is also the inverse
matrix.
C T C CC T I
Cli Clj Cil C jl il
(6-15c)
[Note: We have not used the condition that either eˆi , i 1,3 or eˆi , i 1,3 form a righthanded coordinate system, so orthogonal rotations include both rotations and
transformations consisting of a rotation and a reflection. More about this distinction
later.]
Example: Show that the rotation matrix of (6-10) above is orthogonal.
Solution:
CC
T
cos
sin
0
sin
cos
0
0 cos
0 sin
1 0
cos2 sin 2
sin cos sin cos
0
sin
cos
0
0
0
1
sin cos sin cos
sin 2 cos2
0
0
0 . (6-16)
1
I
Determinant. It is easy to verify the following fact: The determinant of an orthogonal
matrix must be equal to either +1 or -1. (The proof is left to the problems.) If its
determinant is equal to +1, it is said to be a proper orthogonal matrix. All rotation
matrices are proper orthogonal matrices, and vice versa. Rotation matrices must have a
determinant of +1. However, this is not the only condition for a square matrix to be a
rotation matrix. It must also have the property that its transpose is equal to its inverse.
6-4
Vector Spaces in Physics
8/6/2015
NOTE: All rotation matrices are orthogonal; but not all orthogonal matrices are
rotation matrices. Can you explain why?
C. The Rotation Group
Rotations are linear operators which transform three coordinates {xi} as seen in one
coordinate system into three coordinates {x'i} in another system. All vectors transform in
this same way, as given in equation (6-4) above:
A j ' C jl Al
(6-4)
We have shown that C must be orthogonal, and have determinant equal to +1.
Multiplication of two rotation matrices gives another square matrix, and it must also be a
rotation matrix, since carrying out two rotations, one after another, can be represented by
a single rotation. So, the rotation group is defined as follows.
Definition. The "rotation group" consists of the set of real orthogonal 3x3 matrices with
determinant equal to +1.
(a) Group multiplication is just matrix multiplication according to the standard
rules.
(b) The group is closed under multiplication.
(c) The identity matrix is the identity of the group.
(d) For every element A of the group, there exists an element A-1, the inverse of
A, such that AA1 I .
(e) Multiplication satisfies A BC AB C .
These properties are easy to prove and are left to the exercises.
D. Tensors
What are tensors? Tensors look like matrices; but only certain types of matrices are
tensors, as we shall see. They must transform in a certain way under a rotation of the
coordinate system. Vectors, with one index, are tensors of the first rank. Other objects,
such as the rotation matrices themselves, may have more than one index. If they
transform in the right way, they are considered to be higher-order tensors. We will first
discuss rotations, then define tensors more precisely.
Under a rotation of the coordinate system, the components of the displacement vector
x change in the following way:
(6-17)
xi = Rij x j .
The tensor transformation rules represent a generalization of the rule for the
transformation of the displacement vector.
Tensors of Rank Zero. A rank-zero tensor is an object g which is invariant under
rotation:
Rank - zero tensor : g g
(6-17a)
6-5
Vector Spaces in Physics
8/6/2015
Tensors of Rank One. Modeled on the rotation properties of vectors, we define a
tensor of rank one as an object with a single index, C = { C i ,i = 1,3} or just Ci for short,
such that the components Ci' in a rotated coordinate system are given by
(6-18)
First - Rank Tensor : Ci = RijC j .
Tensors of Rank Two. A second-rank tensor is an object with two indices,
D = { Dij ,i, j = 1,3} or Dij for short, such that its components Dij' in a rotated coordinate
system are given by
(6-19)
Second - Rank Tensor:
Dij = Rik R jl Dkl .
That is to say, each of the two indices transforms like a single vector index.
Tensors of Rank Three. Following the same pattern, a third-rank tensor Hijk follows
the transformation law
(6-19)
Third - Rank Tensor : H ijk = R il R jm R kn H lmn .
And so on.
Much of the power of tensor analysis lies in the ease in which new tensors can be
created, with their transformation properties under rotations just determined by the
number of tensor indices. Here are two theorems about creation of tensors which we
offer without proof.
Tensor Products: The direct product of a tensor of rank n and a tensor of rank m
is a tensor of rank n+m.
Thus, for instance, the product of two vectors, AiBj, is a second-rank tensor.
Contraction of Indices: If two tensor indices are set equal (and so summed), the
resulting object is a tensor of rank lower by two than that of the original tensor.
Example: Let us consider the contraction
of the two indices of the tensor AiBj:
(6-20)
Ai Bi A B .
The fact that this dot product is a zero-rank tensor guarantees that it has the same
value in all coordinate systems.
Special Tensors. The Kronecker delta symbol and the Levi-Civita symbol have
indices which look like tensor indices - but their components are defined to be the same
regardless of the coordinate system. Can they still transform like tensors, of rank two and
three, respectively? This is in fact the case; the proof is left for the problems. This
means that these two tensors can be used in conjunction with vectors and other tensors to
make new tensors.
6-6
Vector Spaces in Physics
8/6/2015
Example: Consider the following tensor operations: First form the direct
product ijkAlBm of the Levi-Civita tensor with two vectors A and B. Then
perform two contractions of indices, by setting l=j and m=k. This produces the
following object, with one free index:
l j
ijk Al Bm
ijk A j B k (AxB )i .
(6-21)
mk
This is what we call the cross product or vector product. The theorems about
tensor indices guarantee, since it has one free index, that it does in fact transform
like a vector under rotations.
Example: The trace and determinant of a tensor are both scalars, since all the
free tensor indices are removed by contraction. Thus, the trace and determinant
of a matrix are the same in all coordinate systems, provided that the matrix itself
transforms like a tensor.
E. Coordinate Transformation of an Operator on a Vector Space.
Here is an explanation of the transformation law for a second-rank tensor. Consider an
operator O operating on a vector A to produce another vector B :
(6-22)
B OA .
Tensors are all about looking at something from another point of view - in a rotated
coordinate system. So, if we write quantities in the rotated coordinate system with a
prime, the other point of view is
(6-23)
B OA .
Under rotations, every vector is just multiplied by R . If we apply this to both sides of
equation (6-22) it becomes
(6-24)
RB R OA .
Now insert between O and A a factor of the identity matrix, which we can write as
I R 1 R R T R , where we have used the fact that R's inverse is just its transpose.
RB ROA
ROR T RA
.
(6-25)
ROR T RA
This just shows the operator working in the rotated coordinate system, as shown in
(6.23), provided that the operator in the rotated coordinate system is given by
O ROR T .Similarity Transformation (6-26)
This process, of sandwiching a square matrix between R and its inverse, is called a
similarity transformation.
But . . . shouldn't O transform like a second-rank tensor? Let's write 6-26) in tensor
notation,
6-7
Vector Spaces in Physics
8/6/2015
T
Oij Ril Olm Rmj
Ril Olm R jm .
(6-27)
Ril R jm Olm
This is exactly the transformation law for a second-rank tensor given in eq. (6-19).
This means that laws like
(6-28)
B OA .
can be used in any coordinate system, provided that vectors and operators are
transformed appropriately from one coordinate system to another. This sounds a little
like Einstein's (really Galileo's) postulate about the laws of physics being the same in all
inertial frames. And this is no accident. When laws are written explicitly in terms of
tensors, they are said to be in "covariant form." It makes for beautiful math, and
sometimes beautiful physics.
F. The Conductivity Tensor. There are many simple physical laws relating vectors.
For instance, the version of Ohm's law for a
distributed medium is
z
(6-29)
J E .
E
This law asserts that when an electric field
E is applied to a region of a conducting
material, there is a flux of charge J in the
same direction, proportional to the
magnitude of E . This seems simple and
right. But is it right? ("A physical law
should be as simple as possible -- but no
simpler," Einstein.)
A more general relation can be obtained,
following the principle that J must
x
transform like a vector. We can combine J
Figure 6-3. Graphite, represented as layers (of
in various ways with tensor quantities
hexagonally bonded carbon) stacked one on top
representing the medium, as long as there is
of another.
one free index. The two obvious relations
are
J E J i Ei
scalar conductivity
.
(6-30)
J E J i ij E j tensor conductivity
The scalar relation causes current to flow in the direction of the electric field. But the
tensor relation allows the current to flow off in a different direction. What kind of
medium would permit this?
Graphite is a material which is quite anisotropic, so we might expect a relation of more
complicated directionality. Figure 6-3 shows the model we will take, of planes of carbon
atoms where electrons flow easily, the planes stacked in such a way that current flow
6-8
y
Vector Spaces in Physics
8/6/2015
from one plane to another is more difficult. We can set up a conductivity matrix as
follows. From symmetry, we would expect that an electric field in the x-y plane would
cause a current in the direction of E , with conductivity 0 "in the planes." And, also
from symmetry, an electric field in the z direction should cause current to flow in the z
direction, but with a much smaller conductivity 1. This can be written in matrix form as
0 0 0 E1
(6-31)
J E 0 0 0 E2 .
0 0 E
1 3
If you multiply this out, you see that the first two components of E (the x and y
components) are multiplied by 0, and the third component, by 1.
But what if for some reason it is better to use a rotated set of axes? Then the applied field
and the conductivity tensor both need to be transformed to the new coordinate system,
and their product will produce the current density in the new coordinate system. Note
that scalar quantities, like J E , are zero-th rank tensors and should have the same value
in both coordinate systems.
Numerical Example. In equation (6-31), let's take 1 = 0.1 * 0 (weak current
perpendicular to the graphite planes) and have the electric field in the y-z plane, at an
angle of 45 from the y axis, as shown in figure (6-3):
0
1 0 0
2
E E0
, 0 0 1 0 .
(6-32a)
2
0 0 0.1
2
2
This gives for the current density J in the original coordinate system,
0
0
1 0 0
2
2
J E 0 E0 0 1 0
(6-32b)
0 E0 2 .
0 0 0.1 2
2
2
2
20
The z component of the current is small compared to the y component, and so the current
vector moves closer to the plane of good conduction.
Now let's see what happens in a rotated coordinate system. Let's go to the system where
the electric field is along the z axis. This requires a rotation about the x axis of 45 .
Using the form of this matrix given in eq. (6-11) above, we have
6-9
Vector Spaces in Physics
8/6/2015
1 0
2
R Rx ( 45) 0
2
2
0
2
In this system the electric field vector is simpler:
1 0
0
2
2
E RE E0 0
2
2
2
2
0
2
2
And the conductivity tensor becomes
R RT
1
0 0
0
0
1
2
2
0
2
2
0
2
2
2
2
1 0
0 1
2
2
0 0
0
2
2
2
2
0
0
2
2
0
0
1
0 0 0.55 0.45
0 0.45 0.55
0
0
2
.
2
2
2
(6-33)
0
0
2
E0 0 ;
2
1
2
2
1
0
0 0
2
1 0 0
2
0 0.1
2
0
2
0
0
2
2
2
2
2
2
20 20
(6-34)
0
2
2
2
2
;
(6-35)
Now here is the punch line: We can either rotate J (calculated in the original coordinate
system) into the rotated coordinate system; or we can calculate it directly, using the
electric field and the conductivity tensor expressed in the rotated coordinate system.
Here goes:
6 - 10
Vector Spaces in Physics
8/6/2015
1 0
0 0
0
2
2 2
J RJ 0 E0 0
0 E0 0.45 ;
2
2
2
0.55
2
2 2
0
2
2 20
And the other way:
0
0 0
1
0
J E 0 E0 0 0.55 0.45 0 0 E0 0.45 ;
0 0.45 0.55 1
0.55
It works - same result either way. Here is the interpretation of this example.
(6-36)
(6-36)
The equation
J E
produces the current density flowing in a medium (a vector) in terms of the electric field
in the medium (also a vector). But . . . the current does not necessarily flow in the
direction of the applied electric field. The most general vector result caused by another
vector is given by the operation of a second-rank tensor!
In a simple physics problem, the conductivity tensor will have a simple form, in a
coordinate system which relates to the structure of the material. But what if we want to
know the current flowing in some other coordinate system? Just transform the applied
field and the conductivity tensor to that coordinate system, and use
J E
x
v
G. The Inertia Tensor.
In classical mechanics, a rigid body which is
r
rotating about its center of mass had angular
L
momentum. The rate of rotation at a particular
x
instant of time can be described by a vector, ,
x
which is in the direction of the axis of rotation, with
magnitude equal to the angular speed, in
Figure 6-4. Displacement and velocity,
radians/sec. The angular momentum is also given
and the resulting angular momentum.
by a vector, L , which is equal to the sum (or
integral) of the individual angular momenta of the parts of the rigid body, according to
(6-37)
dL dm r v .
In certain cases, such as when the rotation is about an axis of symmetry of the object, the
relation between L and is simple:
(6-38)
L I .
6 - 11
Vector Spaces in Physics
8/6/2015
Here I is the moment of inertia discussed in a first
course in mechanics.
v r
r
However, surprisingly enough, the resultant total
angular momentum vector L is not in general parallel
to . The general relation involves an angularmomentum tensor, I , rather than the scalar moment
of inertia. The general relation is
L I . (6- Figure 6-5. A rigid body, rotating about
an axis through the origin.
39)
where now, for a body approximated by N discrete point masses,
N
L Li
i 1
and
(6-39)
Li mi ri vi
is the angular momentum of the i-th mass. The velocity of a point in the rotating body is
given by
vi ri ,
and so
dL dm r v
dm r r
We will evaluate the double cross product using tensor notation and the epsilon killer.
dLi dm r r i
dm ijk rj klml rm
dm kij klm rj l rm
(6-41)
dm il jm im jl rj l rm
dm
Problems
The Pauli matrices are special 2x2 matrices associated with the spin of particles like the
electron. They are
0 1
0 i
1 0
, 2
, 3
.
1
1
0
i
0
0
1
(Here “i” represents the square root of -1.)
Problem 6-1. (a) What are the traces of the Pauli matrices?
(b) What are the determinants of the Pauli matrices?
(c) Are the Pauli matrices orthogonal matrices? Test each one
6 - 12
Vector Spaces in Physics
8/6/2015
Problem 6-2. The Pauli matrices are claimed to satisfy the relation
j k k j 2i jkl l
Test this relation, for j=2, k=3.
0 i
Problem 6-3. In Problem 6-1 above one of the Pauli matrices, 2
, did not
i 0
turn out to be orthogonal. For matrices with complex elements a generalized relationship
must be used:
CC † C †C I
C is unitary
†
where the "Hermetian conjugate" matrix C is the complex conjugate of the transpose
matrix.
(a) Calculate the Hermetian conjugate 2† of the second Pauli matrix.
(b) Check and see if it is unitary.
[Unitary matrices become important in quantum mechanics, where one cannot avoid
working with complex numbers.]
Problem 6-5. Using the fact that the determinant of the product of two matrices is the
product of their determinants, show that the determinant of an orthogonal matrix must be
equal to either +1 or -1. Hint: Start with the orthogonality condition AT A I .
Problem 6-6. For each of the three matrices below, say whether or not it is orthogonal,
and whether or not it is a rotation matrix. Justify your answers.
2
2
0
0 0 1
4 3 0
2
2
1
A 0 1 0 , B 3 4 0 , C 0 1
0
5
1 0 0
0 0 5
2
2
0
2
2
Problem 6-7. Consider rotations Rx() and Rz(), with the matrices Rx and Rz as defined
in the text. Here represents the angle of a rotation about the x axis and represents the
angle of a rotation about the z axis.
First, calculate the matrices representing the result of carrying out the two rotations in
different orders:
A Rx Rz
and
B Rz Rx .
Comparing A and B , do you find that the two rotations Rx and Rz commute?
6 - 13
Vector Spaces in Physics
8/6/2015
Next, look at your results in the small-angle approximation (an approximation where
you keep only terms up to first order in the angles and ). Does this change your
conclusion?
Problem 6-8. Show algebraically that the rotation group is closed under multiplication.
that is, show that, for A and B members of the rotation group, the product matrix C = AB
is orthogonal and has determinant equal to +1.
Problem 6-10. The Kronecker delta symbol is defined to have constant components,
independent of the coordinate system. Can this be true? That is to say, is it true that
?
ij Ril R jm lm ij
Use the orthogonality of the transformation matrix Rij to show that this is true.
Problem 6-11. The Levi-Civita symbol ijk, defined previously, has specified constant
values for all of its components, in any frame of reference. How can such an object be a
tensor - that is, how can these constant values be consistent with the transformation
properties of tensor indices? Well, they are! Consider the transformation equation
ijk = Ril R jm R kn lmn ,
where R is a rotation matrix, satisfying the orthogonality condition RijRik = jk, and ijk is
the Levi-Civita symbol. The question is, does ' have the same values for all of its
components as ? As a partial proof (covering 21 of the 27 components), prove that
ijk = 0
if any two of the indices {i,j,k} are equal.
Note: in carrying out this proof, set two indices to the same value - but do not sum over
that index. This operation falls outside of the Einstein summation convention.
Problem 6-12 Tensor transformation of the Levi-Civita symbol. (Challenge
Problem) Prove that under transformations of tensor indices by an orthogonal matrix R,
the Levi-Civita tensor transforms according to the rule
ijk = ijk det( R ) .
Problem 6-13 Invariance of length under rotation. The length of a vector is a scalar,
and so should not change from one coordinate system to another. To check this, use the
vectors E and J from the conductivity example in the text. See if they have the same
magnitude in both coordinate systems used.
Problem 6-14 Invariance of a scalar product under rotation. The inner product of
two vectors is a scalar, and so should not change from one coordinate system to another.
To check this, use the vectors E and J from the numerical example in the discussion of
tensor conductivity in this chapter. See if the inner product E J is the same in both
coordinate systems used.
6 - 14
Vector Spaces in Physics
8/6/2015
Problem 6-15. Return to the numerical example in the discussion of tensor conductivity
in this chapter. Suppose that the electric field E in the original system still has
0
magnitude E0, but is directed along the positive z axis; that is, E 0
E
0
(a) Calculate the current density J E in the unprimed system.
(b) Now consider the same rotation, using R from the example in the text, and calculate
J E in the primed system. (Note: has already been calculated in the notes.)
(c) Now calculate the current density in the primed system by the alternate method,
J RJ . Compare with the result from part (b).
6 - 15