Assigned by
Juan Pablo Fernández
Department of Physics
University of Massachusetts
Amherst, Massachusetts
INTELLIGENT DESIGNS
selected topics in physics
INTELLIGENT DESIGNS
selected topics in physics
x
y
Assigned by
Juan Pablo Fernández
Department of Physics
University of Massachusetts
Amherst, Massachusetts
Selected Topics in Physics: Intelligent Designs
© University of Massachusetts, October 2008
AUTHORS
v
CONTENTS
preface xi
vii
viii contents
P R E FA C E
instructor
Juan Pablo Fernández; 1034 LGRT
jpf@alumni.umass.edu or juanf@physics.umass.edu
teaching assistant
Benjamin Ett
bett@physics.umass.edu
xi
xii contents
evaluation
The grade for this course will be based on five writing projects
assigned in the following order:
1. A journal paper
2. A grant proposal
3. A textbook
4. A science newspaper
5. A final project
In due time I will provide more details about each of the projects
and propose a few different topics from which to choose. If
you would rather write about something else you must tell me
promptly.
You will hand in two drafts of each project. The first one must
be submitted in “draft” form and will receive extensive feedback.
The second draft will be considered final in terms of content and
form and will be assessed as such.
You will have roughly three weeks to complete each project. The
first week you can devote to experimenting with the physics and
the technology, the second to producing the first draft, and the
third to producing the final draft.
At least the first two projects will be peer-reviewed: everybody
will (anonymously) evaluate two papers and have theirs evaluated
by two classmates. The evaluations will include a suggested grade,
usually an integer from 7 to 10, and will themselves be graded.
For the third and fourth projects you will have to collaborate with
your classmates.
The fifth project will be freestyle and may involve media other
than paper.
Part I
ave you ever looked up at the sky and been amazed by its brilliant
H shade of blue? Or why the sun changes color at sunset? In this
chapter we will provide an answer to these questions through the study
of electromagnetic waves. We will uncover why sky is blue and not
white. We will also investigate sunsets and learn why sunsets in some
locations are more beautiful than others.
1.2 waves
2π
T= . (1.2)
kv
Another useful property of waves to know is the frequency of the wave,
which represents the number of oscillations that occur per a unit of
time. The frequency is
1 kv v
f = = = . (1.3)
T 2π λ
Frequency can also be solved for in terms of angular frequency. We can
find angular frequency by
ω = 2π f = kv. (1.4)
3
4 morgan-elise cervo: why is the sky blue?
Figure 1.1: This chart provides wavelengths for the different types of electro-
magnetic waves [9].
∇ · B = 0. (1.7)
The above equation states that the divergence in a magnetic field is zero;
in other words, there are no monopoles in magnetic fields. The third
Maxwell equation is known as Faraday’s induction law and it forms the
basis of electrical inductors and transformers [62]. The equation reads,
∂B
∇×E = − . (1.8)
∂t
In other words, the line integral of the electric field around a closed
loop is equal to the negative of the induced magnetic field [8]. The final
Maxwell equation is a correction to Ampere’s law. It states that in a
static electric field, the integral of the magnetic field around a closed
loop is proportional to the current flowing through the loop. Rewritten
in mathematical terms we have,
∂E
∇ × B = µ 0 J + µ 0 e0 . (1.9)
∂t
1.4 radiation 5
Figure 1.2: Notice how from the horizon the color is gradient of color of longest
wavelength to shortest.
1
B(r, t) = E0 cos(k · r − ωt + δ)(k̂ × n̂). (1.11)
c
where k is the propagation vector, n̂ is the polarization and ω is the
frequency.
1.4 radiation
q+
dz
r
q-
To find the potential vector A, we use the equation for the current in
the wire (the derivative of the z component of charge with respect to
time),
µ0 p0 ω 2 sin Θ
B=− ( cos[ω (t − r/c)]Φ̂. (1.21)
4πc r
So far we have described a wave moving radially outward at a frequency
of ω. We have also defined the electric and magnetic fields, which are
orthogonal to each other and in the same phase. We now find the
intensity of the wave. The energy radiated by an oscillating dipole can
be found by the Poynting vector [36],
1
hSi = ( E × B) (1.22)
µ0
When we use our values for the magnetic and electric field we find,
µ0 p20 ω 4 sin2 θ
hSi = r̂. (1.23)
32π 2 c r2
1.5 blueness of the sky 7
Figure 1.4: Plot of three color cones that shows which wavelengths they best
receive [9]
So why does the sun turn red when the sun is setting? This is
especially puzzling be cause we just explained that red light, with
8 morgan-elise cervo: why is the sky blue?
the lowest frequency, is the least powerful. When the sun is about to
pass over the horizon, the sun’s light has to travel a greater distance
through the atmosphere to reach the observer. The particles in the
earth’s atmosphere cause more waves of shorter wavelengths to be
scattered across this distance. For this reason, the sky is red when the
sun is setting. Similarly, sunsets may be more brilliant in coastal areas
or very dusty area because there are more particles in the air to scatter
light.
1.6 problems
c
n= (2.1)
v
Snell’s Law summarizes an important experimental observation relat-
ing the refractive index of media to the angle that light is refracted. It
says that if the incident angle i and the refracted angle r are measured
with respect to the normal, the ratio of the sines of the angles is equal
to the inverse ratio of the corresponding indexes of refraction. [60].
sin i n
= 1. (2.2)
sin r n2
In the drop of rain the light rays that form the primary bow are twice
refracted and once reflected. The angle D1 between the incident light
ray and the ray that enters the eye to produce the image of a rainbow is
determined by the incident and reflected angles, i and r. More precisely
it is given by
9
10 matthew mirigian: the physics of rainbows
Figure 2.1: This is the path of a ray producing the primary bow. [93]
Figure 2.2: This is the path of a ray producing the secondary bow. [93]
The secondary bow, which can be visible above the primary bow, is
formed by two internal reflections in the drop of rain shown in figure 2.2.
Using the same treatment as for the primary bow we can see the angle
of deviation, D2 to be
can view the peak of the secondary rainbow. Notice in figure 2.3 that
the ray is deviated in the opposite direction as for the primary bow.
This explains the reverse in color distribution compared to the primary
bow.
2.4 dispersion
Figure 2.3: The paths of light forming the primary and secondary bow. [89]
1 sin(r )
=
1.33 sin(30◦ )
r = 22.2◦
2.5 problems
λ0 λ0 630 nm
λw = , n= = = 1.33
n λw 475 nm
Then from n = c/v we can determine the velocity of the light in water.
Figure 3.1: The inverted image is projected on to the opposite wall from the
light coming in through the pin-hole. Image courtesy of [43]
13
14 matthew drake: the camera and how it works
Principles from the camera obscura are still used in modern cameras
which we will explore in detail later in this chapter. The camera obscura,
along with the lenses will be explained in further sections.
3.2 lenses
1 1 1
= + 0 (3.1)
f s s
where f is the focal length, s is the distance of the object to the lens,
and s0 is the distance of the image that is created by the lens, where
all distances are measured along the optical axis. If the image has a
positive distance, then it is a real image. If the image is a negative
distance, then it is a virtual image. Two important properties emerge
from the form of Eq. 3.1.
1. If the object is at the same distance as the focal length, than for
Eq. 3.1 to hold, s10 must go to zero and thus, the image forms at
| ∞ |.
2. If the object is very far away (approaching |∞|), than for Eq. 3.1
to hold, the image forms at the focal length of the lens.
I specify that the distances are at absolute value of ∞ because they can
be towards the positive or negative end of the optical axis. Which end
of the optical axis an image is located on is dependent on whether the
lens is converging or diverging. By definition, if the lens is a converging
lens, than the focal length is positive. If the lens is a diverging lens, than
the focal length is negative. Figure 3.2 shows a typical image formation
from a converging lens. The optical axis has the origin located at the
center of the lens with the left side of the lens typically being the
negative direction.
Along with the placement of the image, we can also examine the size
of the image that is formed. Because the lens bends the light rays, the
size of the image will be different from the size of the object, except in a
special case. The magnification of the image comes from the equation
s0
m=− (3.2)
s
The magnification tells us the ratio of the size of the image to the size of
the object. A negative sign in the magnification tells us that the image
is inverted, as is the case in the camera obscura.
3.3 the camera itself 15
Figure 3.2: The rays from point Q are refracted by the thin lens to bend toward
the focal point. Because the lens is thin, the ray is considered to bend at the
middle of the lens. Image courtesy of [99]
1 1 1
cm−1 + 0 = cm−1 (3.3)
100 s 20
1 5 1 4
= cm−1 − cm−1 = cm−1 (3.4)
s0 100 100 100
s0 = 25 cm (3.5)
25 cm
m=− = −.25 (3.6)
100 cm
The negative magnification tells us that the image is inverted and one fourth
of it’s original size. Since the image distance is positive, we also know that it
is a real image.
Figure 3.3: The image of the key is real, inverted, and smaller than the original
object when it is seen through the camera lens. Image courtesy of [99]
f
f-number = (3.7)
D
where f is the focal length and D is the diameter of the aperture. The
light intensity that may reach the film is proportional to the square of
the inverse f-number, the time that the shutter is open, and also the
brightness of the actual object itself.
Nearly all cameras contain options to zoom in or out on objects. One
method of doing this is to arrange two converging lenses as shown in
Fig. 3.4
The primary lens is a converging lens and brings the light rays
together near it’s focal point and inverted. The image is then seen
through the secondary lens, another converging lens. Because the image
is inverted a second time, it now has the correct orientation and can
be magnified. To obtain the magnification of the zoom lens, you must
3.4 questions 17
Figure 3.4: The secondary lens projects the image of the primary lens to make a
magnified image. Image courtesy of [2]
mtotal = m1 m2 (3.8)
This type of zoom lens is one of the most simple lens and is often
referred to as the “astronomer’s lens.”
3.4 questions
3.5 solutions
19
20 sebastian fischetti: holography: an introduction
(a) (b)
Figure 4.1: Figure (a) shows the hyperbolic interference fringes produced by
two point sources of coherent light, and how these fringes are recorded in a
photographic plate; Figure (b) shows how the interference fringes recorded
by the holographic plate create a virtual image when illuminated by coherent
light [47]. This is an example of a simple transmission hologram.
two sources interfere constructively; this means that the difference in the
path lengths is an integral number of wavelengths: AP − BP = nλ. The
interference fringe passing through P must therefore consist of the locus of
points such that the difference of the distance to A and B is a constant. This is
none other than the geometric definition of a hyperbola; thus the interference
fringes produced by two point sources of coherent light will be hyperbolic
surfaces. In particular, the total number of interference fringes is limited to
the number of wavelengths that can fit into the distance between A and B.
Figure 4.2: Here is how interference fringes are recorded in a reflection holo-
gram; notice that the fringes are approximately parallel to the surface of the
plate [47].
of what’s happening, but there are cases where it fails, at which point we
need to take into account diffraction effects with the zone plate model,
which will be discussed further in Example 4.3.
The thicker the emulsion used to store the interference fringes, the
more well-defined the shape of the fringes, and the more effectively light can
reflect off of them to produce an image. In the case of reflection holograms,
the thickness of the emulsion is especially crucial, because it places an upper
limit to how many Bragg-reflecting surfaces can fit within the emulsion. If the
thickness of the emulsion is significantly greater than the separation between
successive interference fringes, then the hologram is considered thick. If
instead, the emulsion is thinner than the separation between interference
fringes, the hologram is considered thin. A reflection hologram ceases to
exist when it becomes too thin because the Bragg interference is lost. A thin
transmission hologram can still exist, but in this case is called a surface
hologram, since it essentially consists only of surface striations and yields a
lower-quality image than its thick counterpart. A surface hologram cannot be
explained using the geometric model we described, and instead requires
use of the zone plate model described above. In general, thick holograms
are better than thin, as they are capable of containing more information
throughout their depth.
4.6 problems
2. Why would a higher beam ratio be preferable to a low one in the making
of a hologram?
Solution: The beam ratio is meant to reduce intermodulation noise, due
to the interference of the object beam with itself. If the beam ratio is high,
the reference beam’s amplitude is stronger than the object beam’s, and
so the interference of the object beam with itself gives a low amplitude
compared to the high amplitude of the interference of the object beam
and the reference beam, drowning out intermodulation noise.
3. Imagine that in the apparatus of Figure 4.5 we replace the screen to the
right with a transparency containing an array of dots and we replace the
transparency to the left with a screen; then we project a laser beam from
the right. What might we expect to happen?
a) Nothing - the apparatus only works in one direction.
b) An array of letters will be projected on the screen at left, mirroring
the array of dots.
c) It is impossible to tell, since the answer depends on what exactly
the reference hologram is of.
Solution: The correct answer is choice 3b. This is simple symmetry: the
Fourier transform of the array of dots, when convoluted with the Fourier
hologram of the letter, will yield an array of letters.
C O L I N L A L LY: L I S T E N I N G F O R T H E S H A P E O F A
DRUM
5
5.1 introduction
∂2 F
= c2 ∇2 F, (5.1)
∂t2
√
where c is some constant that we shall normalize to be c = 1/2.
There exist special solutions to the wave equation, of the form
This is the one-dimensional limiting case of the general problem being set-up
above
Consider a string of length ` stretched taut between two walls (Figure 5.2).
Its end points are fixed; that is, it obeys the boundary conditions
27
28 colin lally: listening for the shape of a drum
d2 U
+ k2 U = 0, (k2 = 2ω 2 )
dt2
which is just (5.3) in one dimension (minus the boundary conditions included
in 5.3). The general solution to this equation is
Note that since the boundary conditions on F involve only x, they are effec-
tively boundary conditions on U. Applying them, we have
U (0) = 0 =⇒ A=0
nπ
U (`) = 0 =⇒ B sin k` = 0 =⇒ k = ,
`
where n is a positive integer. Thus,
nπ
ω= √
2`
and our normal mode (now labeled by the integer n) is
nπx √
Fn ( x, t) = B sin einπt/ 2` ,
`
where B is a normalization constant.
5.2 eigenvalues
1 2
∇ ψn + λ2n ψn = 0.
2
Naturally, ψn (r) = 0 for any point r that lies on the boundary Γ. Note
that the eigenfunctions are normalized such that
Z
ψn2 (r) d2 r = 1.
Ω
0 l
5.3 problems
1. Carry out the analysis done in the example, but use the boundary
conditions
∂F
F (0, t) = 0, = 0.
∂x x=`
These boundary conditions correspond to the physical scenario
of a string with one fixed end at x = 0, and one end free to move
in the plane at x = `.
30 colin lally: listening for the shape of a drum
M I N D O V E R M AT T E R
CHRISTOPHER MACLELLAN: GLASS 6
6.1 introduction to glass
On the macroscopic scale the differences between a solid and liquid are
obvious. A commonly unknown fact about solids is that there are two
fundamental types: crystalline and amorphous. The distinction between
these types can only be seen on the atomic scale. Crystalline solids are
made up of molecules that form a well defined lattice structure that
is repeated throughout the substance. This is called long-range order,
or translational periodicity. On the other hand amorphous solids have
no long-range order. Atoms in amorphous solids do show connectivity
that resembles the structural units of crystalline states but it is not
regular and does not repeat. This quasi-periodicity is indicative of what
is called short-range order [31]. The difference between the molecular
structures of crystalline solids and amorphous solids is shown in Figure
1.
On the molecular level liquids look exactly like amorphous solids
and as one may expect the molecules in an amorphous solid have a
significant degree of freedom. Then what is the difference between and
amorphous solid and a liquid, and why don’t they actually flow? There
are a few different ways to draw the line between amorphous solids
and liquids. One way is to say that a substance is a solid when the time
required for it to undergo a structural change is much longer than the
33
34 christopher maclellan: glass
Figure 6.1: Atomic Makeup of Crystalline Solids vs. Amorphous Solids [7]
This is not the only thing that can happen when a liquid is cooled. If
cooled fast enough it will not crystallize and will remain in a liquid
state below its melting point. This is called a supercooled liquid and is
also indicated by increasing viscosity as temperature is reduced. Even-
tually the supercooled liquid cools to the point that it forms a glass.
This transition is marked by a glass transition temperature. This tem-
perature is defined as the range in which thermodynamical variables
(e.g. volume, enthalpy, entropy) begin to change. It is a range because
the value of the variables changes in a continuous fashion, unlike the
change that occurs at the melting point. It is important to note that
a substance does not have a specific glass transition temperature. As
mentioned earlier, supercooled liquids only result from sufficiently fast
cooling rates. The cooling rates that will allow a supercooled liquid
to form depends on the substance being cooled. This cooling rate also
changes the region over which a substance can be supercooled. If one
reduces the rate of cooling, the range over which the temperature can
be supercooled increases, which effectively lowers the glass transition
temperature. In fact, the glass transition temperature can vary as much
as 10-20% for very different cooling rates. An odd consequence of these
properties is that the measurement of the glass transition temperature
varies by how the glass is prepared. Despite this knowledge of the glass
transition, there are still many questions left to be answered. Physicists
and chemists still debate the details of the glass transition on a regu-
lar basis and new methods are constantly being formulated to try to
analyze glass structure and transition. [27]
L2
g (r ) =
2π∆rN ( N − 1) ∑ δ (r − |~rik |) , (6.1)
i 6=k
oscillatory nature exists. This indicates that there is some regular change
in density as you get further away from the central particle, which is
what we would expect from a structure with long-range order. This
also agrees with our assertion that crystallinity exists where there is
a relatively monodisperse composition of particles. When z = 3.3mm
we would expect more or less equal mixing between the different size
spheres and an amorphous structure. g (r ) at this height confirms this,
as there is no pattern in how the density of particles changes as r
increases [69].
Using this analysis we have found a way to measure the long range
order in a substance. Although the above treatment is mainly quali-
tative, more analysis techniques are being developed to describe the
positioning of the spheres in both 2D and 3D. [69] Nevertheless, this
method has been proven to be a efficient way of creating system that
is analogous to solids that can be experimentally measured and easily
manipulated. Analyzing colloidal system like the one above provides
one of the best chances for physicists and chemists to unlock the secrets
of how and why glasses form.
2. When cooling a liquid to form a gas, raising the cooling rate the
glass transition temperature
• a)raises
• b)lowers
• c)does not change
38 christopher maclellan: glass
ave you ever noticed how the water droplets that come from
H your faucet splash at the bottom of the sink? Have you ever
watched rain droplets splash as they fall and hit your car windshield
while driving? The phenomenon known as splashing occurs in many
situations such as these. Splashing is also important in technological
and industrial situations such as the ignition of fuel in your car, or when
your printer puts ink onto a piece of paper. This chapter will help you
understand situations such as these by developing an understanding of
the physics of splashing.
Scientists have studied splashing for over a hundred years. Because
of the beauty of the motion involved with splashing, it has been one of
the most highly praised phenomena studied via the use of high speed
photography. From famous sketches and photographs of scientists
such as Harold Edgerton and A.M Worthington[96] , the beauty of
splashing has been available to mainstream society. Yet the physics of
this tremendous phenomenon is still not fully understood.
The goal of this chapter is to develop an understanding of some of
the physics involved when a drop of liquid falls and strikes a smooth
surface. We will see that the there are many factors (characteristics of
our system) that determine how the system will evolve in time. We will
review fundamental concepts such as pressure, surface tension, viscos-
ity, and others to develop a groundwork for understanding splashing
as well.
39
40 robert pierce: the physics of splashing
7.2.1 Pressure
Figure 7.2: Photographs taken by Lei Xu, Wendy W. Zhang, and Sidney R. Nagel
in their experiment involving falling alcohol drops onto a smooth surface. The
top three frames represent alcohonl under an air pressure of 100 kPa, and the
bottom three frames represent alcohol under an air pressure of 30 kPa. The left
frames (top and bottom) are the drop just above the surface, at time t = 0 ms,
the middle frams are the drop at time t = .276 ms, and the right frames are the
drop at time t = .552 ms. We see that in the middle frames a corona or liquid
layer is spreading out from the center, and by the time expressed in the right
frames, the stress due to air pressure has won, and we see splashing.
7.2 pressure, surface tension, and other concepts 41
PMg
ρA = (7.3)
kb T
We will see that after a drop strikes the surface, a liquid layer will
spread outwards (the corona), and interact with the atmosphere. There
will be stress on the liquid layer due to the atmosphere that will be
proportional to, among other things, the density of the gas in the
atmosphere. We will learn what the other things are later on, but
first make sure that you are capable of using the above equations to
analyze this system in terms of the atmospherically applied stress.
Make sure that you understand pressure in the context of our system
by performing the following example.
1. When observing a liquid drop spread out after coming into contact
with a smooth surface, we notice that the liquid spreads out very
rapidly. Under these circumstances the stress applied to the spreading
liquid layer due to the atmosphere will be proportional to the density
of the gas in the atmosphere. If our atmosphere is made of air, and
if air has a molecular mass of M29 = 29u (u = 1.66x10−23 kg), what is
the pressure that the atmosphere may exert on a liquid layer that is
spreading outwards? Assume that we observe this at room temperature
(T = 295K). The density of air is measured to be ρ A = 1.2kg/m3 ,
and Boltzmann’s constant is k b = 1.381x10−23 m2 kgs−2 K −1 . Does this
pressure look familiar?
Solution: Here we may use 7.2 to solve the problem. We are given values
for the variables ρ A , Mg , and T. Solving for pressure, we find
ρ 1.2kg/m3
P= kb T → P = (1.381x10− 23m2 kgs− 2K − 1)(295K ).
Mg 29 ∗ (1.66x10− 23kg)
(7.4)
liquid section itself, and the middle layers will move with velocities in
between the two extremes. Let’s make shear stress more clear: stress
is defined as the average force applied to a surface per unit area (see
equation 7.6), and shear means that the stress is being applied parallel
to the surface. Shear stress will result as layers move towards or away
from other layers. Stress τ is proportional to force F by
F
τ= . (7.6)
A
Where A is area. Shear stress is directed parallel to the liquid. Ultimately
we will see that the stress in the liquid is acting to hold the liquid
together. We will see that this stress will interact with the stress on the
liquid from the atmosphere (due to pressure as described in earlier
sections), and when the atmospheric stress is stronger, splashing will
occur.
First, let us devote some time to viscosity. In our study of viscosities
we will be interested in what is known as kinematic viscosity. This is
when we look at the relationship between the resistive force in a liquid
due to viscosity, and the inertial force of the liquid. Kinematic viscosity
νL of a liquid is defined as
µ
νL = , (7.7)
ρ
where t is the time elapsed since the drop struck the surface.
Now we are in a position to consider the stress on the expanding
liquid layer due to the liquid layer itself. The stress on the boundary
liquid layer will be due to the surface tension of the liquid striving to
keep the layer intact. With d defined as above, we have that the surface
tensional stress (shear stress) is
σ σ
ΣL = → ΣL = √ (7.9)
d νL t
7.3 splashing
Splashing in our system will occur when the stress on the liquid layer
due to the atmosphere is stronger than the stress on the liquid layer
due to the liquid. This is because the stress due to the atmosphere is
trying to break the liquid layer apart, and the stress due to the liquid is
trying to keep the liquid layer together. We’ll have two equations for the
7.4 summary 43
two different stresses. Earlier we discussed how the stress due to the
atmosphere should be dependent, among other things, on the density
of the air. We also know that the stress is proportional to the rate of
expansion of the liquid layer, and the speed of sound in the atmosphere.
If Σ A is the stress due to the atmosphere, then our equation for the
stress is
s r
PMg γk b T RV0
Σ A = (ρ A )(C A )(vl ) −→ Σ A = . (7.10)
kb T Mg 2t
√σ
s √
ΣA νL t
q RV0 νL
= = γMg P (7.12)
ΣL
q
2k b T σ
q
PMg γk b T RV0
kb T Mg 2t
From this ratio we can see that a more viscous liquid will splash
more easily than a less viscous liquid. This is counter intuitive because
one would think that a more viscous liquid would stay together more
easily, however, it is apparent that a more viscous liquid will make the
value of the ratio in (7.12) larger, which implies that the liquid layer
is more likely to break apart and splash. So splashing is clearly not a
straightforward phenomenon at all.
7.4 summary
1 For a derivation of this equation, see the paper Drop Splashing on a Dry Smooth Surface by
Lei Xu, Wendy W. Zhang, and Sidney R. Nagel, 2005.
44 robert pierce: the physics of splashing
2. Consider the answer found in the previous question. If the shear stress in
the expanding liquid due to the liquid is Σ L = 4500 kPa when observing
for time t = .5 ms, will there be any splashing? Explain, why or why
not?
Solution: From the first question we have that the stress on the expanding
liquid layer due to the atmosphere is
Σ A = 4903.8 kPa. (7.18)
We know that if the stress due to the atmosphere is greater than the
stress due to the liquid, then the expanding liquid layer will break apart,
and we will see splashing. So the ratio from equation 7.12 will be greater
than 1 when we have splashing, and less than 1 when we don’t have
splashing. In our case,
ΣA 4903.8 kPa
= = 1.09 > 1 (7.19)
ΣL 4500 kPa
We do see splashing. Again, this is because of the fact that the stress due
to the atmosphere (the stress that is trying to break the expanding liquid
layer apart), is stronger than the stress due to the liquid itself (the stress
that is trying to hold the liquid together). Ultimately the atmosphere
wins, the liquid breaks apart, and we see splashing.
7.6 multiple choice questions 45
n the deep ocean it is common for mariners to see wave heights that
I are in the 5-meter to 12-meter range. In the worst storm conditions
waves generally peak at 15 meters (49 ft). Commercial ships have been
built to accommodate waves of such heights. About once a week a ship
at sea sinks to the bottom of the ocean [4]. Over the last two decades
more than two dozen cargo ships measuring over 200 meters in length
have been lost at sea [38]. Most of the time these ships will sink without
a mayday signal or any trace left behind of its existence.
For centuries mariners have described waves that appear out of
nowhere that approach 35 meters (115 ft.) in height. However there was
no scientific evidence to back up this anecdotal evidence. It is important
to distinguish that these types of waves are described to appear out of
nowhere in the deep sea. This freakish wave is drastically different than
that of the tsunami. A tsunami is created by a disturbance in the ocean.
The most basic example of this is the example of a stone being dropped
in water, which creates waves that are generated radially outward.
Likewise, a quick shift of plates making up the sea floor creates a series
of low waves with long periods. Disturbances known to create tsunamis
range from seismic activity, explosive volcanism, submarine landslides,
or even meteorite impacts with the ocean [13]. Tsunamis occur near
land as their name suggests; tsun translates from Japanese in harbor
and ami into wave (The phenomena known was originally known as a
tidal wave, however, this name incorrectly implies that they are related
to the tides).
While tsunamis are much more deadly than the freak waves de-
scribed by mariners (The megatsunami of December 26, 2004, killed
over 250,000 with waves in the 10-to-30 meter range along the coast of
Indonesia, India, Thailand, Sri Lanka, and elsewhere), their behavior is
much more understood as about three fourth’s of tsunamis originate
from submarine seismic disturbances. These disturbances produce a
great amount of energy that translates into the waves as they travel at
speeds approaching 200m/s with wavelengths on the order of hundreds
of kilometers [67]. Tsunamis are not felt at the deep sea as they only
produce wave heights around a meter or two, which is not detectable
with the typical ocean swell. However, once the waves created by the
disturbance approach areas of shallow water, more energy is packed
into the waves, creating waves in the tens of meters along the shore. In
the 20th century over 1000 tsunamis have been recorded with 138 being
labeled as a “damaging” tsunamis [12]. Since the last half of the 20th
century oceanographers have been able to identify the instigating dis-
turbance for the occurrence of a particular tsunami. This is completely
different than freak waves out of nowhere described by mariners which
carried the same type of lore and about as much comprehension as
mythical monsters like Loch Ness.
For a long time oceanographers dismissed the possibility of the exis-
tence of these types of waves with such a high rate of occurrence. What
47
48 sam bingham: freak waves
Figure 8.1: The crest heights during a 20-minute interval at the Draupner Oil
platform at 15:20 GMT. The peak is 18.6 meters from the surface.
is known as the linear model has been used to predict wave heights out
at the deep sea with great accuracy. The commercial shipping industry
has based most of the designs for its ships off of this model because of
the accuracy of the linear model. The freak waves seen by the mariners
are part of the natural wave spectrum but should only occur about
once every ten thousand years according to the accepted linear model.
Clearly something was wrong.
While oceanographers had started to put more faith and energy into
the belief of such freak waves in the 1970’s the major breakthrough
came with the Draupner wave on January 1st, 1995 at the Draupner
oil platform in the North Sea. The Draupner oil platform, which is
located directly west of Norway, had a laser-based wave sensor at
an unmanned platform that would record waves in a fashion that
would be uninterrupted by the platform’s legs [41]. It should be noted
that oceanographers base much of their analysis on what is called
the significant wave height, or swh, which is the average height from
trough to crest of one third of the largest waves during a given period.
On January 1st at the Draupner Oil platform the swh was in the 11-to-
12-meter range. The maximum wave height measured that day was 26
meters and the peak elevation was 18.5 meters, as shown in figure 8.1.
The annual probability for a wave of that height in the 12 meter swh is
10−2 , or once in a hundred years. While the causes for such waves are
understood with much more confidence since the Draupner wave, it is
important to determine what waves are freak waves by nature.
Now that we understand the Draupner wave and a bit about these
freak waves let’s take a look at just how odd these waves are when
using the linear model. In the most simple model of ocean waves,
the sea elevation is taken to be a summation of sinusoidal waves of
different frequencies with random phases and amplitudes. In the linear
approximation the wave field is taken as a stationary random Gaussian
process. The probability density distribution is described as
1 2 /2σ2
f( η ) = √ e−η . (8.1)
2πσ
8.3 interference 49
In Eq. 8.1, η is the elevation of the sea surface with a zero mean level,
hη i = 0, and σ2 is the variance computed from the frequency of the
spectrum, S(ω )
Z
σ2 = h η 2 i = S(ω )dω. (8.2)
2 /8ω 2
P( H ) = e− H . (8.3)
Equation 8.3 gives the probability that the wave heights will exceed a
certain height, H. We can now introduce the significant wave height,
Hs , into the equations and begin to see its relevance. Extensive work by
Stanislaw Massel has shown that Hs ≈ 4σ[58]. From this we can find a
direct relation into the probability for a certain wave height to occur
under the conditions that are present by rewriting Eq. 8.3 as
2 /H 2
P( H ) = e−2H s . (8.4)
Could a rogue wave occur in shallow water of only 20 meters in depth with
the top third average waves only around 2 meters in height? If one could
what would be the most probable rogue wave and with what probability
would it occur?
Solution: This problem is quite simple but requires careful thought on the
definition we established for a rogue wave. We stated that a rogue wave is a
wave with a wave height of twice that of the significant wave height, thus a
5-meter wave would be a rogue wave if the swh was 2 but not if it was 4. So if
we return to the question we can surely have a rogue wave in shallow water
with a swh of 2 meters if the wave height is greater than 4. This 4-meter wave
would also be the most probable rogue wave with a probability that is found
2 2
from Eq. 8.4. With our conditions we get P(4) = e−2(4) /2 = 3.5 × 10−4 .
These types of waves are referred to as shallow water rogue waves and are
just as freakish as the giant ones in the deep sea. The highest recorded ratio
H/HS was a wave height of 6m in conditions with a significant wave height
of 2m.
8.3 interference
While the Draupner wave was a tremendous breakthrough for the field
of oceanography, as it proved that such waves do exist and that there
50 sam bingham: freak waves
were inaccuracies in the standard linear model, it did not show how
or why such waves form. In order to examine the possible causes for
these waves researchers looked at the areas in which they occurred. By
plotting areas in which freak wave phenomena were described in the
past a pattern was found. A great number of the waves occur in areas
where there are fast currents that run counter to the wave direction. The
most extreme example of this is the Agulhas current off of the Cape
of Good Hope and the tip of South Africa. The Agulhas current is the
fastest current on the planet. Figure 8.3 shows 14 reported freak waves
found in a study by Captain J. K. Mallory prior to 1975. His study
found similar results in the Gulfstream current, the Kuroshio current
and any others in the Pacific Ocean.
Figure 8.2: These are the results of a study by Captain Mallory done in 1976.
The circles represent cases where abnormal waves were observed, the dotted
line is the continental shelf, and the other depths show the continental slope.
The abrupt change in depth is what creates the Agulhas current and gives it its
jet like function parallel to the shoreline [49].
Traveling around the tip of South Africa has always been a rough
stretch for mariners, but the area of Agulhas current created waves that
mariners said could reach 30 meters in wave height. In this region the
waves are generated off Antarctica and move northward towards South
Africa. They run unopposed until they meet the Agulhas current, which
moves in a southwest direction parallel to the east coast of South Africa.
Because of this area of interaction the wind waves are made up of two
different systems: the long swell waves coming from the Antarctica
and short/steep waves generated by the sea wind of South Africa. The
superposition of these two systems will undoubtedly contribute to an
increase in wave height due to possible constructive interference, but
not enough to account for the rogue waves reported in the area.
The solution to these types of waves relates to geometric optics. If
we think about a strong variational current as a lens then it could
focus opposing waves and create points where enough energy would
converge to explain the freak waves. This would be possible because
the areas of the strong flow of the current would retard the wave
advancement much more than the weaker flow, which would cause the
wave crests to bend. This refraction allows for the focusing of the waves
energy and allows the waves to grow in height as shown in figure 8.3.
8.3 interference 51
Variational Current
b
Wave
crests
Wave
orthogonals
Figure 8.3: a shows the interaction of the two systems that are present off the tip
of South Africa [49]. b shows the refraction that waves undergo when flowing
against a current that has a variable surface speed. This refraction can lead
to the build up of rogue waves when the wave orthogonals converge [Figure
made in Inkscape].
52 sam bingham: freak waves
∂A ∂A ω ∂2 A ω0 k20
i( + c gr )= 0 + | A|2 A, (8.5)
∂t ∂x 8k0 ∂x2 2
and the surface elevation, η ( x, t) is given by
1
η ( x, t) = ( A( x, t)ei(k0 x−ω0 t) + c.c. + ...). (8.6)
2
Here k0 and ω0 are the wave number and frequency of the carrier
wave, c.c. denotes the complex conjugate, (...) determines the weak
highest harmonics of the carrier wave, and A is the complex wave
amplitude, which is a function of x and t. At this point what is known as
the Benjamin-Feir instability comes into play. It is very well known that a
uniform wave train with an amplitude of A0 is unstable to the Benjamin-
Feir modulational instability corresponding to long disturbances of
wave number, ∆k, of the wave envelope satisfying the relation
∆k √
< 2 2k0 A0 . (8.7)
k0
The highest instability will occur at ∆k/k0 = 2k0 A0 which gives a
maximum growth a rate of ω0 (k0 A0 )2 /2 [48]. There has been a great
deal of research on the nonlinear stage of the Benjamin Feir instability
analytically, numerically, and experimentally (see Grue and Trulsen
(2006) and Kharif (2003)). Through this research it is apparent that wave
groups can appear or disappear on the order of 1/[ω0 (k0 A0 )2 ] on the
timescale. This behavior can be explained by breather solutions of the
non-linear Schrödinger equation. Figure 8.4 shows the formation of
waves that have very high energies and are created by the Benjamin
Feir instability. There are many different breather solutions that can
create various large waves. What is referred to as the algebraic solution
has a maximal wave height of 3 times the waves around it and some
solutions such as the Ma-breather and the Akhmediev breather are
periodic in time and space.
8.5 conclusion 53
Figure 8.4: The evolution of weakly modulated wave train (numbers - time
normalized by the fundamental wave period). Highly energetic wave group
formed due to the Benjamin Feir instability [48]
8.5 conclusion
One area that has received the least amount of research to date comes
from the focusing potential of wind. Experimental research in this
area has shown that with out wind waves will focus at a fixed are
and create a rather high amplitude wave but with strong wind the
high amplitude wave will be further downstream and with a higher
amplitude. Experimental results has also showed that rogue waves can
occur under the conditions of a strong wind when they typically would
not if the wind was not present. This line of research has lead most to
believe that strong winds can increase the rate of occurrence of rogue
waves. The effect the wind has is to weaken the defocussing process
which leads to a greater chance for the rogue wave. As a result of this
finding it makes sense that rogue waves are more likely to occur when
storms with vicious winds are present.
54 sam bingham: freak waves
8.6 problems
F
n v
w
Ff
Figure 9.1: The basic Amontons-Coulomb model of friction, ~Ff = µ~F.
55
56 paul hughes: friction
m
n
v
w
θ
tan(θ ) = µ (9.3)
a.
b.
c.
where sgn(v) = |v|/v. This model assumes that what we call “rough-
ness” on the macroscopic scale appears, microscopically, as “asperities”
or projections from the mean plane of each surface [29].
One way in which this model can break down is in the case of
microscopically smooth, geometrically complementary surfaces: the
microscopic contact area reaches its maximum with very little surface
deformation, meaning that an increased normal force will no longer
increase shearing stress or friction. In such a case, various atomic forces
(such as covalent bonds, electrostatic attraction, and “Van der Waals”
forces between dipoles) are also particularly likely to play a role: as
a noteworthy example, the energy required to separate the interface
surfaces is equal to the free energy associated with their bonding, minus
the elastic potential energy made available as elastic deformation is
relaxed [28]. Obviously, then, a limit on elastic deformation implies
greater influence of adhesion forces as surfaces become smoother, and
when adhesive forces become involved, the friction quickly becomes less
Coulombic–that is, the relationship between friction force and normal
force becomes less linear–as a dependence on the area of interface is
introduced.
In the same way, lubricants serve to fill gaps and minimize the defor-
mation necessary for asperities to bypass one another by functioning
somewhat like ball bearings: asperities are permitted to “coast” past
one another, rather than striking, pressing, and producing a deforma-
tion on the scale of their geometric overlap—the friction gains, instead,
a dependence on the lubricant’s viscosity. Unsurprisingly, then based
on this conceptual model, the highest coefficients of friction are seen
58 paul hughes: friction
in materials which are either pourous and highly elastic (e.g., rubber),
or which otherwise exhibit a highly elastic ‘clinging’ action (e.g., vel-
cro). Interfaces such as a highly porous, elastic rubber on hard, rough
concrete can yield coefficients of friction well above µ = 1; others in-
volving such materials as teflon or ice (especially near the melting point,
where the contact surface can melt and self-lubricate) can drop below
µ = 0.04.
9.4 summary
9.5 problems
conveyor belt?
We define the initial velocity vblock of the block relative to the
conveyor belt as vblock (t = 0) = −v.
Thus Ff = µn = µmg cos(θ ), opposed by mg sin(θ )
Ftotal = mg[µ cos(θ ) − sin(θ )] so a = g[µ cos(θ ) − sin(θ )]
The box may accelerate to match the conveyor belt’s velocity, in
which case:
v1 = v0 + at, where v1 = v and v0 = 0, so tstop = g[µ cos(θv)−sin(θ )]
Conversely, the box may still be sliding relative to the conveyor
belt when it reaches the end:
s1 = s0 + v0 t + 12 at2 where s0 = v0 = 0 and s0 = `, so tedge =
q
2`
g[µ cos(θ )−sin(θ )]
Whichever happens first is the time at which the block stops slid-
ing and reaches its peak velocity.
π
b. If m = 1kg, µ = 0.75, v = 20m/s, θ = 12 , and ` = 2m, at
what horizontal distance d from the edge of the conveyor belt will
the the block q
land?
4
First, tedge = 9.8[0.75 cos(π/12 )−sin(π/12)]
= 0.936s,
20
and tstop = 9.8[0.75 cos(π/12 )−sin(π/12)]
= 4.38s.
Therefore, we see that the block leaves the conveyor belt at
t = 0.936s with velocity
vedge = at = g[µ cos(θ ) − sin(θ )]t = 4.27m/s and at height
hedge = ` sin(θ ) = 0.518m
From here, the problem is simply one of projectile motion:
h gnd = 0 = hedge + vedge sin(θ )t − 12 gt2 so tland = 0.231s
d = vedge cos(θ )tland = 0.953m.
61
62 keith fratus: neutrino oscillations in the standard model
of spin along the z axis being either along the positive or negative z
axis [37].
We can represent the two eigenstates as vectors in a complex-valued
Hilbert Space:
! !
1 0
; . (10.2)
0 1
The theory makes a restriction on the form of the operators that can
represent observables, which is that they must be hermitian. Because
of this fact, all possible eigenvalues must be real, and the eigenstates
of the operator must form a basis of all possible states that the particle
can attain. In our example of spin, this means that the particle’s z-
component of intrinsic angular momentum can in general be a linear
combination of the two basis states, or,
! ! !
a 1 0
= a + b , (10.3)
b 0 1
where a and b are some given complex numbers. But the question now
arises, what value will we get when we measure the z-component of
spin, for the particle in this generalized state? The answer is that we
could get either of the two eigenvalues, with some given probability.
The probability of returning each eigenvalue is given by the square of
the norm of the coefficient on the associated eigenstate. This, in essence,
is the principle of superposition in quantum mechanics [37].
Because there are often many physical properties that a particle can
possess, there will generally be many different hermitian operators
that can act on the state of a particle. Each of these operators will thus
define a new basis in which we can define the state of our particle,
corresponding to the eigenstates of this operator.
Figure 10.1: The Standard Model of elementary particle physics. Image courtesy
Fermilab.
The fermions come in two types, the quarks, which are in the first
two rows, and the leptons, which are in the last two rows, with one
of the biggest differences among them being that the leptons do not
interact with the strong force. In reality, matter particles can form
composites which have integer spin and are thus bosons, and the
particular bosons represented in the standard model are specifically
referred to as gauge bosons. But in terms of fundamental particles, we
can safely make the distinction between fermions and bosons without
too much consequence. This general analogy of passing bosons back
and forth to mediate a force can be extended to the rest of the standard
model. For each interaction in the standard model, there is a set of
mathematical rules describing how it acts on certain particles, and one
or more bosons that mediate it. In addition to the strong force, there is
the familiar electromagnetic force, mediated by the photon, and also
the weak force, something we will discuss in more detail later. One of
the more peculiar aspects of the standard model is that the strong force
actually grows stronger with distance, as if the quarks were attached by
springs. The result of this is that a free quark can never be observed, and
that the existence of quarks must be inferred by indirect means. There
are some unresolved issues in the standard model, some of which hint
at physics beyond the standard model, but aside from this, the standard
model, for the most part, still represents our current understanding of
fundamental physics [88].
The origin of these particles is described by something called “Quan-
tum Field Theory.” Quantum Field Theory (or QFT) generally states
that each of the fundamental particles is sort of a “bump” in something
called a quantum field, that can move through space. For example,
there is an electron field that extends throughout space, and actual
electrons are like ripples moving across the “surface” of that field (of
course, this visual analogy can only be pushed so far). Speaking some-
64 keith fratus: neutrino oscillations in the standard model
what more mathematically, the fields that exist throughout space have
certain normal modes of oscillation, and when energy is input into
these fields, the resulting excitations of the field are what we experience
as fundamental particles [14, 78].
Furthermore, QFT associates with each of these fields something
called a “creation operator” (there are also annihilation operators, but
never mind that for now). Many students have encountered these sorts
of operators without ever having been aware of it. A common problem
in introductory quantum mechanics is to solve for the possible energy
eigenstates of the simple harmonic oscillator. One method for solving
this problem is to use lowering and raising “ladder” operators, which
create lower and higher energy states from an existing one. When we
take an eigenstate of the simple harmonic oscillator, and apply the
lowering operator to it, we attain the eigenstate with one less unit
of energy. What we are actually modeling when we do this is the
creation of a photon that is emitted from the system, as it carries
away the difference of the energies of the two states. In QFT, these
operators “create” all of the possible particles. A given operator acts
on the vacuum of space, and “creates” an instance of the particle
associated with that operator. For example, the operator associated with
the electron field “creates” instances of electrons. When a photon turns
into an electron-positron pair, the creation operators for the electron
and positron model this process (along with the annihilation operator
for the photon) [37, 14].
While all of this may seem like useless information, seeing as how we
do not intend to pursue all of the associated mathematical details of
Quantum Field Theory, there is an important caveat to all of this. As we
mentioned before, there are often many different operators associated
with a particle. This is also true in QFT. Looking at Figure 10.1, we
see a type of boson called a W particle, which is responsible for medi-
ating an interaction called the weak force. In essence, the weak force
is responsible for transforming one type of particle into another. The
W boson actually represents two different particles, the W − and W +
particles, which are named according to their electric charge. Because
of this, a particle must lose or gain a unit of charge when emitting a W
particle (remember that forces are generally mediated by the exchange
of bosons). Since total electric charge must always be conserved, and
because every particle has an intrinsic value of electric charge, this
means that one type of particle must transform into another type of
particle when emitting a W boson. There is actually another type of
particle associated with the weak force, the Z0 boson, but it is involved
in a slightly different interaction. It is responsible for the scattering
of one particle off of another, similar to how charged particles can
scatter off of one another via the electromagnetic interaction. This re-
semblance to electromagnetic phenomena is actually due to a very deep
connection between the electromagnetic and weak forces. Electroweak
theory states that above a certain energy threshold, electromagnetic and
weak interactions become indistinguishable, and the two interactions
are “unified.” Of course, this subject could easily form the basis of an
entirely separate chapter, but for now we can safely concentrate on just
10.4 the weak and higgs mechanisms 65
pair of quarks, since the weak interaction will take care of how the
second particle from each pair is affected by the phenomenon. Let’s
work with the down and strange quarks, and call them d’ and s’ when
referring to their representation by the weak force, and d and s when
discussing them in terms of the Higgs mechanism. Quantum theory
says that these two representations should be related to each other by a
sort of “change of basis” matrix in some given Hilbert space in which
we describe the state of our physical system, which will look something
like
! ! !
d cos θ − sin θ d0
= . (10.4)
s sin θ cos θ s0
The angle θ is a measure of how much of a “rotation” there is between
the two bases. If we expand this relation, we get
d = cos θ d0 − sin θ s0 (10.5a)
But back to the issue of the weak and Higgs mechanisms. It is often
customary, as mentioned previously, to call d and s “mass states,” and
d’ and s’ “weak states.” Each of them can be represented as linear
combinations of the others, and are related by the equation given
previously. This analogy actually extends to other particles. In particular,
we will consider the neutrinos, which are in the bottom row of Figure
10.1. Recent evidence suggests that they behave in a manner similar
to quarks, in that their mass states and weak states are different (for
a while it was believed that neutrinos were massless, and hence were
not subject to this behavior, but this has been shown to not be true).
Neutrinos are fermions, and there are three known types of them. They
each come paired with another fermion, one that is either the electron
(in the case of the electron neutrino), or one that is similar to the electron
(either the muon or tau particle, being identical to the electron, except
for mass) [14].
So what is the significance of this property of neutrinos? It actually
just so happens that this property of neutrinos is responsible for some-
thing called neutrino oscillations. That is, it is possible for one type
of neutrino to turn into another type of neutrino as it travels through
space. This can also occur for quarks, but because of the fact that a
free quark can never be observed, it involves the composite particles
made up of quarks, so the details are somewhat simpler for the case of
neutrinos. The case of neutrinos is also particularly relevant, not only
because this phenomenon occurs to a greater extent among neutrinos
than it does for quarks, but because the existence of neutrino oscilla-
tions implies that they interact with the Higgs mechanism, and thus
indeed have mass, something that was not believed to be true for several
decades. Neutrino oscillations are also significant in the sense that they
are of great experimental interest. Our current models that describe
the nuclear activity in the sun predict that we should be detecting a
greater number of electron neutrinos being emitted from it than we
actually are. It is believed that the explanation for this discrepancy is
that of neutrino oscillations; that is, electron neutrinos emitted from the
sun have the ability to change into other types as they travel between
the sun and earth. There are of course many other reasons why the
phenomenon is of great interest, but suffice it to say, it is a behavior
that beautifully demonstrates some of the basic properties of quantum
mechanics, without too much mathematical complexity [14].
Having discussed the significance of these “oscillations,” let’s see if
we can get a basic mathematical derivation of them. For the sake of
simplicity, we’ll consider the case of neutrino oscillations among the
first two families. That is, oscillation among the electron and muon
neutrinos. Once again, this is a somewhat reasonable approximation to
the more complex case of three-family mixing. Because each neutrino
comes paired with an electron-type particle by the weak force, we can
consider the mixing matrix with respect to just one particle in each of
the pairs. In other words, we can write the mixing equation with respect
to just the neutrinos. If we denote the electron and muon neutrinos with
respect to the Higgs mechanism as e and µ, respectively, and denote
68 keith fratus: neutrino oscillations in the standard model
where θ is the angle that describes the “rotation” among the repre-
sentations (note that we are now referring to the weak states as the
result of the mass states being rotated through some angle, instead
of the other way around; this is possible because rotations are always
invertible). Let’s consider a muon neutrino that is created through the
weak force. Because it is created through the weak force, and the weak
force operates with respect to weak states, the neutrino can definitely
be said to be a muon neutrino with respect to the weak mechanism. In
other words, it is in a pure weak state. This means that it can be given
by a linear combination of mass states, or,
where Ee and Eµ are the energies of the electron and muon neutrino
mass states, respectively. We use the letter ν now, because as our state
evolves over time, it will not necessarily be a muon neutrino, or even
an electron neutrino; in general it will be some linear combination of
the two.
What we are interested in, of course, is the probability that we will
detect the original neutrino as being an electron neutrino some time
after it is created. To do this, we follow the standard prescription of
quantum mechanics, which says that the probability of measuring an
electron neutrino is the square of the norm of the projection of our
state at a given time along the basis vector representing the electron
neutrino. Now, when me detect a neutrino, we do so by studying the
way it interacts with matter via the weak force. So any neutrino we
detect will be in a definite weak state, since the weak force always acts
on weak states. This means we need to calculate the projection of our
state onto the weak state of the electron neutrino. Symbolically, we can
write this as
If we take this inner product in the basis of the mass states, then
mathematically, our expression for the inner product becomes
!
0 − sin θ e−iEe t/h̄
he | ν(t)i = ( cos θ , sin θ ) . (10.11)
cos θ e−iEµ t/h̄
10.5 the origin of neutrino oscillations 69
he0 | ν(t)i = − cos θ sin θ e−iEe t/h̄ + sin θ cos θ e−iEµ t/h̄ , (10.12)
1
P( µ0 → e0 ) = ( sin 2θ )2 ( −e−iEe t/h̄ + e−iEµ t/h̄ ) ( −eiEe t/h̄ + eiEµ t/h̄ ).
2
(10.14)
1
P( µ0 → e0 ) = sin2 2θ ( e0 + e0 − (ei(Ee − Eµ )t/h̄ + e−i(Ee − Eµ )t/h̄ )),
4
(10.15)
∆Et
1
P( µ0 → e0 ) = sin2 2θ 2 − 2 cos , (10.16)
4 h̄
where ∆E is the difference between the two energies, and we have used
the definition of the cosine function in terms of complex exponentials.
If we factor out the two, and use the trigonometric identity
χ
2 sin2 = 1 − cos χ, (10.17)
2
then this ultimately simplifies to
∆Et
P( µ0 → e0 ) = sin2 2θ sin2 (10.18)
2h̄
where p is the momentum of the object, c is the speed of light, and m is the
mass of the object. If we factor out the first term in the square root, we can
write
21
m2 c2
E = pc 1+ 2 . (10.20)
p
Using a Taylor expansion, the approximation
1 x
(1 + x ) 2 ≈ 1 + (10.21)
2
allows us to write the energy as being approximately
m2 c2
E ≈ pc 1 + , (10.22)
2p2
or,
m2 c3
E ≈ pc + . (10.23)
2p
This approximation is reasonable, because our neutrinos will generally be
moving relativistically, and this implies that the energy attributable to the
mass of the particle is small compared to the kinetic energy of the particle,
and the second term in equation 10.20 is small enough for the approximation
to be valid.
If we now consider the energy difference of the two mass states, we can
use the above approximation to write
m2e c3 m2µ c3
∆E = Ee − Eµ ≈ pc + − pc − , (10.24)
2p 2p
which simplifies to
c3
∆E ≈ (m2 − m2µ ), (10.25)
2p e
or,
c3
∆E ≈ (∆m2 ), (10.26)
2p
10.6 implications of the existence of neutrino oscillations 71
0.9
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
0 5 10 15
Distance (m) 4
x 10
where ∆m2 is the difference of the squares of the two different masses.
We have assumed that the momentum of the two mass states is the same,
which is a somewhat reasonable approximation. Because the neutrinos move
relativistically, we can assume that the distance x traveled by a neutrino is
related to the time after creation by
x
t≈ . (10.27)
c
If we now use the approximations given in equations 10.26 and 10.27, our
expression for the probability of measuring an electron neutrino is
c ∆m2 x
2
P( µ0 → e0 ) ≈ sin2 2θ sin2 . (10.28)
4ph̄
Once again using a relativistic approximation, we can say that the overall
energy of the neutrino, Eν , is roughly given by
Eν ≈ pc. (10.29)
There are several reasons why our final expression in example 10.1 is
of great interest to experimentalists. First, notice that the oscillations
explicitly depend on the difference of the squares of the two masses. If
the difference were to be zero, then our expression would be identically
72 keith fratus: neutrino oscillations in the standard model
zero for all time, implying that there would be a zero probability to mea-
sure an electron neutrino. In other words, there would be no oscillation.
Because neutrino oscillations have now been experimentally verified,
we know that neutrinos must indeed have mass (since the masses can-
not all be zero and different from each other at the same time), which
has tremendous implications for several branches of physics, from high
energy physics to cosmology.
The fact that neutrinos have mass shows a similarity between quarks
and leptons. Because quarks and (some) leptons were known to have
mass, it was known that quarks and (some) leptons must interact with
the Higgs mechanism in some way. Thus, it was surprising that only
quarks exhibited oscillation, and not leptons. Now that leptons have
been shown to exhibit oscillation, it shows a similarity between the
two groups of fermions. This is crucial for “Grand Unified Theories,”
theories which hypothesize that above some threshold energy, all of
the interactions in the standard model unify into one fundamental
interaction that acts amongst all of the fermions in the standard model
on an equal footing. The fact that we see slight differences amongst the
different fermions, but in general they have the same behavior, suggests
a sort of “broken symmetry” that would presumably be restored above
this energy threshold. This energy threshold (referred to as the Planck
energy) however is believed to be around 1028 eV, which is much higher
than the energies that can be reached in modern accelerators (the Large
Hadron Collider at CERN will only be able to reach energies of 14
TeV) [76, 14].
The fact that neutrinos have mass is also of interest to cosmologists,
since it could have an influence on the rate of expansion of the universe.
Neutrinos are generally created quite frequently in nuclear processes, so
a large number of them flood the universe. Because of this, the fact that
they have mass has serious implications for the energy-mass density of
the universe, which, according to the theory of general relativity, affects
the manner in which the fabric of space-time deforms and evolves with
time [76].
The first real evidence for neutrino oscillations was provided by the
Sudbury Neutrino Observatory in Sudbury, Ontario, Canada, which
detected neutrinos from the sun. While previous experiments had
strongly suggested the existence of neutrino oscillations, the experi-
ment in Canada was the first one to definitively identify the number
of neutrinos of each type, and the experimentally determined num-
bers agreed with theoretical predictions. Previous experiments had
only shown a lack of electron neutrinos being detected from the sun,
and gave no information regarding the number of neutrinos of other
types [14].
Amazingly, without resorting to any of the more advanced mathemat-
ical tools of Quantum Field Theory, it is possible to get a quantitative
understanding of a quantum mechanical effect that has enormous
implications for modern research in physics. Some of the homework
problems investigate the subject with specific numerical examples, to
give a better idea of some of the numerical values actually involved
in neutrino oscillations. Of course, any reader who is interested is
encouraged to research the subject further on his or her own, since
Quantum Field Theory is a very complex, intellectually rewarding sub-
ject that should begin to be accessible to students towards the end of
their undergraduate study.
10.7 problems 73
10.7 problems
where, as usual, d and s are the mass states, and d’ and s’ are the
weak states. Explain why this might account for the the rate of
neutron decay being less than expected. The mismatch between
these two types of quarks also accounts for the existence of an-
other type of decay. A lambda particle is one that is comprised
of an up, down, and strange quark. One of the ways it can decay
is to a proton, where the strange quark turns into an up quark.
This is called lambda beta decay. One might expect this decay
type to be prohibited, since the up and strange quark are in two
different families. Explain how this decay might be possible, and
how it is related to the previously mentioned issue of neutron
beta decay [76].
Solution: The key to understanding this problem is realizing that, as
mentioned before, the weak force only acts on weak states, and that the
weak force is the mechanism that drives the two decay types. Each of the
mass states in the quark composites is a linear combination of two weak
states, so each has some probability of being measured as one of the two
weak states. In the neutron, the down quark that we expect to decay into
the up quark is actually a linear combination of the two weak states, and
the probability that it will be a down quark weak state is proportional to
the square of the coefficient in the linear combination that is applied to
the down quark weak state. This coefficient is the cosine of the rotation
angle, which implies that squaring this term will give us the probability
that we will select the down quark weak state. Of course, this implies
that there is some probability that with respect to the weak force, we
will have a strange weak state, which is not allowed to decay into an
up quark. So the probability that the neutron will not be able to decay
into a proton is given by the square of the sine of the rotation angle. So
the decreased rate is essentially explained by the fact that in some sense,
the down quark mass state in the neutron is not always a down quark
with respect to the weak mechanism, and so the neutron will not always
decay to the proton.
This also explains how a lambda particle can decay into a proton. The
strange quark in the lambda particle likewise has some probability of
74 keith fratus: neutrino oscillations in the standard model
being a down quark with respect to the weak force, given by the square
of the sine of the rotation angle. So there is a finite probability that the
down quark weak state component of the strange quark mass state will
interact with the weak force, allowing the lambda particle to decay into a
proton.
The rate of neutron beta decay will be proportional to the probability
that the down quark mass state in the neutron is in the weak down quark
state, which is given by the square of the cosine of the rotation angle.
The rate of lambda beta decay will be proportional to the probability
that the strange quark mass state in the lambda particle is in the weak
down quark state, which is given by the square of the sine of the rotation
angle. The sum of these two probabilities of course gives unity, so in
some sense, the rate of decay of the lambda particle “makes up for” the
missing rate of neutron beta decay.
Experimentally, the rotation angle between the first two families is found
to be about 0.22 radians, for a decreased rate of neutron decay of about
four percent [76].
2. For the final expression derived in example 10.1, find the spatial wave-
length of oscillation.
Solution: The wavelength of the square of the sine function will be the
first value of the distance that returns the sinusoidal function to zero,
since the square of the sine function oscillates back and forth between
zero and one. The sine of any integer multiple of π will be equal to zero,
so if λ is the wavelength of oscillation, and x is the distance from the
source, then the argument that will be equal to π when the distance x
from the source is equal to the wavelength is simply πx/λ. Equating this
to the expression for the argument found in the example, we have
πx c4 ∆m2 x
= . (10.32)
λ 4h̄cEν
Rearranging this, we have
4πh̄cEν
λ= (10.33)
c4 ∆m2
for the spatial wavelength of oscillation. Note that a larger energy im-
plies a longer oscillation wavelength, since the neutrino will travel a
longer distance in the amount of time it takes to oscillate from one type
of neutrino to another. A larger difference in the masses will cause a
smaller wavelength, since the rate of oscillation increases with larger
mass difference. Note that the wavelength of oscillation is not affected by
the rotation angle among the two families of neutrinos.
3. Using the result derived in the example, find the probability that the
emitted neutrino will be measured as an electron neutrino one hundred
kilometers from the source if it has an energy of one million electron volts.
The experimentally determined value for the mixing among the first two
families is 0.59 radians, and the value found for the difference between
the squares of the masses is 8 · 10−5 eV 2 / c4 (see Figure 10.2) [14].
Solution: If we square the sine of twice the rotation angle, then the value
we attain is
If we substitute the value of the constants into the expression for the
argument to the second sinusoidal function, we have
2
c4 ∆m2 x (c4 ) · (8 · 10−5 eVc4
) · (105 m)
= −
, (10.35)
4h̄cEν (4) · (197.327 · 10 eV · m) · (106 eV )
9
10.8 multiple choice test problems 75
where we look up the value of h̄c. All of the other constants cancel, and
the value of the argument becomes roughly 10.1355. Thus, the probability
is approximately
Note that because of the value of the first sinusoidal term, the maximum
probability to measure an electron neutrino at any given time is roughly
eighty-five percent.
I N F O R M AT I O N I S P O W E R
A N D Y O ’ D O N N E L L : FA S T F O U R I E R T R A N S F O R M 11
11.1 introduction
Due to the orthogonality of the trig functions, it can be seen that only
only at a certain values of 2π f will values of the integral survive.
The equation for the inverse continuous Fourier Transform is given
by
Z ∞
x (t) = X ( f )e j2π f t dt. (11.3)
−∞
79
80 andy o’donnell: fast fourier transform
N −1
X (m) = ∑ x (n)e− j2πnm/N . (11.4)
n =0
N −1
1
x (n) =
N ∑ X (m)e j2πnm/N . (11.5)
n =0
figure(1)
hold on;
clf;
title('Fourier Transform of sin(t)')
xlabel('Frequency Hz')
ylabel('Magnitude of X(m)');
plot(index,Y,'.-')
figure(2)
hold on;
plot(index,y)
title('sin(t)')
xlabel('Time in seconds');
ylabel('x(n)')
As seen from the figures of the signal and the Fourier Transform of it, there
is only one position on the frequency space where we find it. This is backed
up by out expectations about it.
11.4 the fast fourier transform 81
sin(t)
1
0.8
0.6
0.4
0.2
x(n)
−0.2
−0.4
−0.6
−0.8
−1
0 0.2 0.4 0.6 0.8 1
Time in seconds
Figure 11.1: One period of a sine function.
As you can see above, for the Discrete Fourier Transform to be useful,
there needs to be a faster way to computer it. For example, if you were
to have a value of N = 2, 097, 152, using the Fast Fourier Transform
algorithm it would take your computer about 10 seconds to do it while
using the Discrete Fourier Transform described above it would take
over three weeks. [3] This next section will get into the details and
derivation of the radix-2 FFT algorithm. First, we start off with our
original definition of the discrete Fourier Transform. [46]
N −1
X (m) = ∑ x (n)e− j2πnm/N . (11.6)
n =0
2.5
Magnitude of X(m)
1.5
0.5
−0.5
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
Frequency Hz
Figure 11.2: The FFT of a sine function. Notice how it only has discrete compo-
nents .
82 andy o’donnell: fast fourier transform
For the next part of this, we split the sums into evens and odds and
get
( N/2)−1 ( N/2)−1
X (m) = ∑ x (2n)e− j2π (2n)m/N + ∑ x (2n + 1)e− j2π (2n+1)m/N .
n =0 n =0
(11.7)
In the second term, we can easily pull out the phase angle. Doing that
we find
( N/2)−1 ( N/2)−1
X (m) = ∑ x (2n)e− j2π (2n)m/N + e− j2πm/N ∑ x (2n + 1)e− j2π (2n)m/N .
n =0 n =0
(11.8)
Next we need to define some notation to simplify the exponential
terms. We shall now use the following notation.
WN = e− j2π/N , WN
n
= e− j2πn/N , WN
2
= e− j2π2/N , WN
nm
= e− j2πnm/N
(11.9)
Using this notation, we can then replace the above equation with
( N/2)−1 ( N/2)−1
X (m) = ∑ 2nm
x (2n)WN m
+ WN ∑ 2nm
x (2n + 1)WN . (11.10)
n =0 n =0
e− j2π/( N/2)
= WN/2 . This allows us to change the equations to
( N/2)−1 ( N/2)−1
X (m) = ∑ nm
x (2n)WN/2 m
+ WN ∑ nm
x (2n + 1)WN/2 . (11.11)
n =0 n =0
( N/2)−1 ( N/2)−1
∑ ∑
n(m+ N/2) m+ N/2 n(m+ N/2)
X (m + N/2) = x (2n)WN/2 + WN x (2n + 1)( N/2) − 1WN/2 .
n =0 n =0
(11.12)
Next we can use the following expression that
n(m+ N/2)
WN/2 nm
= WN/2 nN/2
Wn/2 nm
= WN/2 (e− j2πn2N/2N ) = WN/2
nm nm
(1) = WN/2 .
(11.13)
We then call the express in front of the summation the twiddle factor
and we can simplify the above as
( N/2)−1 ( N/2)−1
∑ ∑
(nm) m nm
X (m + N/2) = x (2n)WN/2 − WN x (2n + 1)WN/2 .
n =0 n =0
11.5 multiple choice 83
(11.15)
N −1
Xsum (m) = ∑ (x1 (n) + x2 (n))e j2πnm/N . (11.17)
n =0
N −1 N −1
= ∑ (x1 (n))e j2πnm/N + ∑ (x2 (n))e j2πnm/N . (11.18)
n =0 n =0
[46]
T I M M O R T S O L F : T H E P H Y S I C S O F D ATA S T O R A G E 12
12.1 introduction
M = 2N . (12.1)
85
86 tim mortsolf: the physics of data storage
N = log2 M. (12.2)
Solution Since there are only four different values of the DNA base,
12.3 storage capacity is everything 87
we can simply use formula 12.2 to solve for the number of bits that are
required
There are several different schemes that we could use encode these bits. Here
is one that encodes the bases in alphabetical order of the symbols.
Binary/Base-2 Number DNA base
00 A (adenine)
01 C (cytosine)
10 G (guanine)
11 T (thymine)
A byte has eight bits of information. Since each DNA base requires
2 bits for encoding, then a byte can store the value of 8/2 = 4 DNA bases in
a sequence. To represent sequences longer than four DNA bases we would
simply use additional bytes.
∆E∆t ≥ h̄.
88 tim mortsolf: the physics of data storage
Figure 12.1: Transfer rates of magnetic disk drives and microscopic materials.
The transfer rates of single bits of information on a magnetic drive has continued
to improve as the density as increased, but we may be approaching a limit
a trade off is required. The datapoint for the silicon atom shows the transfer
rate of gold atoms bonded to a silicon surface that were read with an scanning
tunneling microscope (STM). The data rate depicted for DNA is presumably
how fast the DNA is able to be transcribed into an RNA molecule. The rate of
RNA transcription is typically greater than 50 RNA bases per second. (Courtesy
of [42])
Hard disk drives (HDDs) are the major data storage devices used by
today’s desktop computer for permanent storage of data. Although
HDDs were introduced into the market place in the 1956 by IBM, the
12.4 the physics of a hard disk drive 89
Figure 12.2: Diagram of the major internal view of a hard disk drive. (Courtesy
of [63])
earlier desktop PCs used removable floppy disk drives (FDDs) because
of the high cost and relatively low data storage of HDDs available
at the time. During the early 1990s, HDDs quickly supplanted FDDs
in desktop PCs as technology permitted HDDs to be made at lower
cost and with higher storage capacity. In this section we will explain
how hard disk drives work and the physics of giant magnetoresistance
(GMR), which is a newer technology that has enabled modern HDDs
to have data storage capacities of 1 Terabyte.
Read/write heads
HDD read/write heads read data from and write data onto a hard disk
platter. Each platter surface has its own read/write head. A disk drive
with five platters has ten heads since there are each side of a platter has
a magnetic surface. The heads are attached to a single actuator shaft
that positions the heads over a specific location of the platter. The drive
controller is the electronic circuitry that controls the rotation of the
drive and the position of the heads. The computer interfaces with the
drive controller to read and write data to the drive. The computer refers
to a specific location on a hard drive by using a head number, cylinder
number, and sector number. This drive controller uses the head number
to determine which platter surface the data is on, the cylinder number
to identify what track it is on, and the sector number to identify a
specific 512 byte section of data within the track. To access the data, the
drive controller instructs the actuator to move to the requested cylinder.
The actuator moves all of the heads in unison to the cylinder, placing
each head directly over the track identified by the cylinder number.
The drive controller then requests the head to read or write data as the
sector spins past the read/write head.
When the hard disk drive is turned off, the heads rest on the platter
surface. When the drive is powered up and begins to spin, the air
pressure from the spinning platters lift the heads slightly to create a
very tiny gap between the heads and the platter surface. A hard drive
crash occurs when the head hits the surface while the platter is spinning
and scratches the magnetic media.
Magnetic surface
The magnetic surface, or media layer, of a hard disk platter is only
a few millionths of an inch thick. The earlier hard drive models used
iron oxide media on the platter surfaces because of its low cost, ease
for manufacturing, and ability to maintain a strong magnetic field.
However, iron oxide materials are no longer used in today’s HDDs
because of its low density storage capacity. Increases in storage density
require smaller magnetic fields so that the heads only pick up the
magnetic signal of the surface directly beneath them.
Modern disk drives have a thin-film media surface. First a base
surface called the underlayer is placed onto the platter using hard
metal alloys such as NiCr. The magnetic media layer is deposited on
12.4 the physics of a hard disk drive 91
Figure 12.3: Depiction of the magnetic bit orientation and read head signals
of media surfaces encoding using LMR (longitudinal magnetic recording) and
perpendicular magnetic recording (PMR) technology (Courtesy of [95])
information on the surface are smaller. In the future, all HDDs will
probably use PMR until a newer technology comes along to replace it.
Figure 12.4: Graph of the GMR effect of the resistance on thin magnetic alloy
layers separated by a thin nonmagnetic spacer. (Courtesy of [85])
12.5 summary
Digital computers and data storage systems store numbers using binary
bits because it is practical to build devices that have two states, on and
off. A byte made by combining the values of eight bits into a single
quantity and can represent 28 = 256 different values. Data storage ca-
pacity and transfer rates have been increasing since the beginning of the
digital era, but as we scale the devices down to the atomic level, there
will be a trade off between these two metrics. This limit is explained
by the Heisenberg relation which relates the amount of time it takes to
force a physical system into a state to the energy of the state the system
is forced into, ∆E∆t ≥ h̄. Hard disk drives (HDDs) are the major data
storage devices in use today. They store information by magnetically
encoded physical bits of information onto a hard disk platter that is
coated with a thin-layer media surface composed of magnetic alloys.
HDDs read information from the magnetic surface from a change in
electrical resistance that occurs from magnetoresistance. Giant magne-
toresistance (GMR) is a recent technology that shows greater response
to changes in magnetic fields by designing the read heads to have thin
layers of magnetic alloys separated by a thin nonmagnetic spacer.
12.6 exercises
2. The first generation of hard disk drives that were produced, relied
on the magnetoresistance effect to decode the magnetically stored
bits of data?
(a) True
(b) False
Solution (a)
4. The GMR effect that has been used to increase hard drive density
is primarily used to:
(a) Decode bits of information on the magnetic thin-layer media
(b) Encode bits of information on the magnetic thin-layer media
(c) Decode and encode bits of information on the magnetic thin-
layer media
Solution (a)
T I M M O R T S O L F : T H E P H Y S I C S O F I N F O R M AT I O N
T H E O RY
13
13.1 introduction
97
98 tim mortsolf: the physics of information theory
The answer to a question that can assume the two values ’yes’ or ’no’
(without taking into account the meaning of the question) contains one unit
of information.
Since almost all of our coin flips are "heads", the sequence of coin flips
will be a long series of "heads" with an occasional "tail" interspersed to
break the series. Our encoding scheme takes advantage of the biased
nature by recording only those places in the sequence where a "tail" flip
has occurred. Since there are 1,000,000 coin flips, each "tail" requires a
20-bit number to record its position in the sequence. Out of 1,000,000
coin flips, the average number of tails will be 1, 000, 000/100 = 10, 000.
On average it will take "10, 000 × 20 = 200, 000 bits to record the results
of each experiment. Thus, by using a very simple encoding scheme we
have been able to reduce the data storage of our experiment by a factor
of 1/5.
For our initial recording scheme that recorded the values of each coin
flip as a single bit, the amount of data required to store the message
was always 1,000,000 bits, no more and no less. With our new encoding
scheme we can only discuss the statistical distribution of what the
amount of data required for each experiment will be. If we look at the
almost impossibly unlucky result where each of 1,000,000 coin flips
is "tails" that occurs for (100)1,000,000 of our experiments, the encoding
scheme would use "1, 000, 000 × 20 = 20, 000, 000 bits of data. This is
100 times larger than our average value of 200,000 bits that we expect
to use for each trial. These improbable unlucky scenarios are balanced
by scenarios where we get lucky and have much fewer coin flips with
"tails" than the average result and can encode our experiment in much
less than 200,000 bits of data. The important thing to note is that almost
all of experiments will require nearly 200,000 bits of data storage, but
the exact value will be different for each result.
We can design an even better encoding scheme than this one. Instead
of recording the absolute sequence number of each coin flip with a
"tails" result, we could just record the difference between the absolute
sequence number of each coin flip of "tails". For example, if a coin flips
of "tails" occurred on the coin flip values
{99 - 0 = 99, 203 - 99 = 104, 298 - 203 = 95, 302 - 298 = 96, 390 - 302 = 88}.
(13.2)
For this encoding scheme, we use 8-bit to encode differences less than
256, while our original compression scheme used 20-bit numbers for
those cases where the "tails" result to save its absolute position in the
sequence. With our latest encoding scheme, on average it will take
"10, 000 × 8 = 80, 000 bits to record the results of each experiment. But
can we do better than this? As you have probably guessed, we can
indeed. The limit to how well we can do can be calculated by Shannon’s
general formula for information uncertainty that we will develop in the
next section.
M
H= ∑ Pi ui . (13.10)
i
M
H = − ∑ Pi log2 Pi Shannon’s formula (bits per symbol). (13.11)
i
S = k B ln W (13.12)
M = W, (13.13)
S = k B ln M = k B ln 2 I = k B I ln 2, (13.14)
S
I= (13.15)
k B ln 2
This formula (13.15) relates the information content we can store in the
device to the thermodynamic entropy of the physical system. Unfor-
tunately, there is not a simple analytical formula that we can use to
compute the entropy of an arbitrary device from knowledge of physi-
cal properties such as mass, volume, temperature, and pressure. The
computation of entropy requires complete knowledge of the quantum
energy levels available to each particle in the device. One author used
a method to approximate a lower bound for the maximum amount of
entropy in a 1 kg device that occupies a volume of 1 L and assumes
that most of the energy of the system arises from blackbody photon
radiation [52]. Using these simplifications, he arrives at a value of
T = 5.872 × 108 K for the temperature at which the maximum entropy
occurs, and S = 2.042 × 108 J K−1 for the total entropy. Substituting
these values into our equation for the information content,
18 months 1 year
time required = 60.6 times to double × ×
1 time to double 12 months
time required ≈ 91 years
Thus it will take about 90 years to develop the technology to store data at our
arrived value of the maximum theoretical limit for data storage. Although
this calculation does have several simplifying assumptions, it does show that
there is a lot of room left to increase the data density relative to a device that
precisely encodes information content into the entropic states of a physical
system.
13.5 summary
13.6 exercises
1. Your assistant Bob rolls a six-sided dice and you don’t know the
result. What is the minimum number of yes/no questions that
you could ask Bob to get the answer? Try to use information
theory to answer this question.
Solution Before posing a set of questions, let’s first use informa-
tion theory to try to encode the result. The dice and have six
results and using a binary coding scheme, we need three bits to
encode the results. There are several different schemes that we
could use encode the dice result. Here is one that encodes them
in order of lower to higher values.
104 tim mortsolf: the physics of information theory
2. Repeat the same problem but this time Bob rolls two six-sided
dice and you only want to determine the value of the sum.
Solution The sum of two six-sided dice rolls will have values from
2 (when two 1’s are rolled) through 12 (when two 6’s are rolled).
Our encoding scheme needs to encode eleven values which takes
four bits. However, since in this case some results are more likely
than others, let’s see if we can design a more clever encoding
scheme.
Binary/Base-2 Number Sum of dice Probability
1
0000 2 36
2
0001 3 36
3
0010 4 36
4
0011 5 36
5
0100 6 36
0101 6
6
0110 7 36
0111 7
1000 7
1001 7
5
1010 8 36
1011 8
4
1100 9 36
3
1101 10 36
2
1110 11 36
1
1111 12 36
13.7 multiple choice questions 105
14.1 introduction
Ant trail formation and general large scale organized travel is made pos-
sible by the indirect communication between individual ants through
the environment. The individual ants deposit pheromones while walk-
ing. As an ant travels, it will probabilistically follow the path richest in
pheromones [22]. This allows ants to adapt to changes or sudden ob-
stacles in their path, and find the new shortest path to their destination.
This incredibly powerful social system allows entire colonies of ants to
make adaptive decisions based solely on local information permitting
ants to transport food far distances back to their nest.
The organizing process, stigmergy, is also used in a similar way
by termites. When termites build mounds they start by one termite
retrieves a grain of soil and sets it in place with a sticky glue-like saliva
that contains pheromones. In this case the mark referred to in the literal
meaning of stigmergy is the grain of soil. This pheromone filled saliva
then attracts other termites who then glue another grain of soil on
top of the last. As this process continues the attractive signal increases
as the amount of pheromones increase [86]. This process is called,
stigmertic building. Stigmertic Building allows the termites to construct
pillars and roofed galleries. The simplest form of stigmertic building is
pillar construction. As the pile of grains of soil accumulate upwards it
naturally forms a pillar. Once the pillar reaches a height approximately
equal to that of a termite standing on its hind legs the termites begin
107
108 sam boone: analytical investigation of the optimal traffic organization of socia
The stigmergy concept in general has led to simple algorithms that have
been applied to the field of combinatorial optimization problems, which
includes routing in communication networks, shortest path problems
and material flow on factory floors [66]. While these applications can
aid efforts to make transportation routes become more efficient other
considerations must be made, such as, congestion and bottlenecking.
This problem applies to user traffic on the internet as well as motor
vehicle traffic.
Ant Colony Optimization(ACO) based algorithms have also begun to
be used in multi-objective optimization problems. In recent years video
game designers have started to use ACO based algorithms to manage
Artificial Intelligence in their games. In particular, these algorithms
are used to manage functions like multi-agent patrolling systems. In
order for this function to run properly, all agents must coordinate their
actions so that they can most efficiently visit the areas of most relevance
as frequently as possible. Similar algorithms have been used in robotics,
computer network management and vehicle routing. Their has also
been recent break-throughs in using ACO to create artificial 3D terrain
objects for video games, as well as engineering, simulations, training
environments, movies and artistic applications.
ACO algorithms can also be very useful for site layout optimization.
This proves to be very valuable, whether applied to the construction
of highways and parking garages to manufacturing plants and ports.
Any construction site layout that is concerned with the positioning and
timing of temporary objects, machines and facilities that are used to
carry out the actual construction process, as well as the permanent
objects being created with regards to their ultimate purpose that will
be carried out, will greatly benefit from the solutions of these prob-
14.2 ant colony optimization 109
lems. Ultimately these ACO based algorithms will greatly enhance the
efficiency while decreasing the cost of these sites.
While ACO research is still very young and fairly scarce, the number
of possible applications seem to be limitless. Almost any field of science,
mathematics, engineering, computer design, finance, economics and
electronics can be greatly improved and expedited with the application
of ACO.
x + y = 12. (14.1)
y = 12 − x. (14.2)
P = xy2 , (14.3)
P = x (12 − x )2 . (14.4)
110 sam boone: analytical investigation of the optimal traffic organization of socia
dx (12 − x )2
P0 = = −2x (12 − x )2 + (12 − x )2 , (14.5)
dx
which simplifies to become,
P( x = 0, y = 12) = 0, (14.7)
P( x = 4, y = 8) = 256, (14.8)
P( x = 8, y = 4) = 128, (14.9)
P( x = 12, y = 0) = 0. (14.10)
x = 4, y = 8, (14.11)
P = 256. (14.12)
One of the famous optimization problems that the Ant Colony Opti-
mization Meta-Heuristic has been applied to is the Traveling Salesman
Problem. The Traveling Salesman Problem is the problem to find the
shortest closed path in which one could visit all the cities in a certain set.
While this algorithm or others of the same nature could be tweaked to
handle this problem with different variables we will concern ourselves
with a Traveling Salesman Problem in which there exists a path (edge)
between any pair of cities.
When a colony of ants forage for food, the ants will initially leave their
nest and head for the foraging area in which the food that they are
looking for lies. Lets assume that their is a natural path from the nest
to the foraging area which twice forks and reconnects. Therefore their
are four possible paths to the food in which the ants can travel. At
first, there is no information on these paths that could give the ants
14.4 applying aco meta-heuristic to the traveling salesman problem 111
Figure 14.3: Double bridge experiment. (a) The ants start to randomly choose
paths to foraging area. (b) The majority of the ants eventually choose the
shortest path (Graph) The distribution of the percentage of ants that chose the
shortest path in the double bridge experiment [24]
While we will use Marco Dorigo’s artificial ants to solve our optimiza-
tion problem hands-on we must recall that all of our artificial ants will
be acting based upon natural ant behaviors that have been witnessed
in nature. For this problem we will use three ideas that have been
witnessed in nature by real ants [23]:
Our artificial ants will move from city to city on the Traveling Salesman
Problem (TSP) Graph based on a probabilistic function that incorporates
both heuristic values that depend on the length of edges (length of
trails between cities) and trail accumulated on edges [23]. The artificial
112 sam boone: analytical investigation of the optimal traffic organization of socia
ants will prefer cities rich in pheromone trail and which are connected
by edges (which in our case is all possible cities). Every time step when
an ant travels to another city they modify the pheromone trail on the
edge being used. Dorigo calls this local trail updating. Initially we will
place m artificial ants randomly across our (TSP) Graph at randomly
selected cities. Once all of our ants have completed a tour the ant that
made the shortest journey will modify the edges used in his tour by
adding pheromone amounting to a total that is inversely proportional
to the tour length. This step is dubbed global trail updating.
As we stated earlier, the artificial ants that we are using act based upon
observed behaviors of real ants in nature, however these artificial ants
were given some capabilities that the real ants do not possess. How-
ever, all of these differences are necessary adaptations that must take
place in order for the artificial ants to be able to discovery solutions
to optimization problems in their new environment (the TSP Graph).
The fundamental difference between the artificial ants used in the ACO
algorithm and the biological ants their design is based upon is in the
way that they modify the trails they travel. Like the ants observed in
nature who deposit pheromones on the surface of their environment,
the artificial ants modify their surroundings by changing some numeri-
cal information that is stored locally in the problem’s state [24]. This
information includes the ant’s current history and performance which
can be read and altered by ants who later visit that region. This informa-
tion is called artificial pheromone trail. This artificial pheromone trail is the
only means of communication between each individual member of the
colony. In most ACO algorithms an evaporation mechanism is installed,
much like real pheromone evaporation, that weakens pheromone infor-
mation over time allowing the ant colony to gradually forget its past
history so that it can modify its direction without being hindered too
much by its past decisions [24].
There are a few traits that the artificial ants possess that the real ants
do not [24].
1. The artificial ants possess an internal state that contains the mem-
ory of its past actions.
2. The artificial ants deposit an amount of pheromone that correlates
to the quality of the solution.
3. The artificial ants timing of when they lay their pheromone does
not reflect any pattern of actual ants. For example, in our case
of the Traveling Salesman Problem the artificial ants update the
pheromone trails after they have found the best solution.
4. The artificial ants in the ACO algorithms can be given extra capa-
bilities by design like lookahead, backtracking, local optimization
and others, that make the overall system much more efficient.
that are not apart of that ants working memory Mk . The ant chooses
the next city s using the probabilistic formula:
∈ Mk [ τ (r, u )] · [ η (r, u )] ,
β
s = argmaxu/ (14.13)
s = S, (14.14)
if s ∈
/ Mk . Otherwise,
pk (r, s) = 0, (14.16)
3
V0 = (32 − x2 ) (14.25)
4
√
Therefore, when V 0 = 0, x must be equal to +/ − 32. SInce x is
a measurement it must be positive. Therefore,
√
x = 32 f t. = 5.66 f t., (14.26)
and
y = 2.83 f t. (14.27)
The volume of the box is then,
V = 90.56 f t2 . (14.28)
117
118 christopher kerrigan: the physics of social insects
colony as the place where more ants are produced. This has an obvious
analogy to the battery (or EMF) in an electric circuit. Also, we can safely
say that the constrained group in which the ants move is probably a
single-file line because of the width of ant tunnels (and from general
observation.) Knowing that the processes of both systems occur in a
similar fashion, perhaps we can use the equations of electric circuits to
describe the motion of the ants.
N
I= , (15.1)
t
where N is the number of ants. With this simple and obvious compar-
ison, we can draw many conclusions about the motion of a group of
social insects.
Answer:
15.6 differences
this logic, the ant who has taken the shorter path will return to the nest
faster (Fig. 4). This means that just outside the nest, the pheromone
density is twice as high on the shorter path (because it has been passed
twice, whereas the longer path has only been passed once until the ant
makes it back to the nest). Other ants leaving the nest, then, will opt for
the shorter path because of the higher pheremone density. Over many
iterations, the ants further reinforce the shorter path and after a while
this will be the only path used (Fig. 5).
15.8 problems
1 1
P[ F (2)] = ( ) (15.3)
2 2
1
P[ F (n)] = ( )n (15.4)
2
2. Suppose the distance between two points a and b is d ab and Cab is the
pheromone density along this line. What is the probability for a single ant
122 christopher kerrigan: the physics of social insects
Cab
P= (15.5)
d ab
Part IV
W H AT ’ S O U T T H E R E
R O B E RT D E E G A N : T H E D I S C O V E RY O F N E P T U N E 16
16.1 introduction
Before we can look into exactly how it was that astronomers found
Neptune, we must first understand the laws that govern celestial bodies.
The interaction of any two masses was described by Newton with his
Law of Universal Gravitation which states that every point mass acts on
every other point mass with an attractive force along the line connecting
the points. This force is proportional to the product of the two masses
and inversely proportional to the square of the distance between the
points
m1 m2
F=G (16.1)
r2
where the force is in Newtons and G is the gravitational constant equal
to 6.674 × 10−11 N m2 kg−2 [101].
Ms Mu
Fs−u = G = 1.399 × 1021 N (16.2)
( R u )2
Mn Mu
Fn−u = G = 2.224 × 1017 N (16.3)
( Rnu )2
So the force acting on Uranus due to the Sun’s gravity is 4 orders of magnitude
larger than the force due to Neptune’s gravity.
125
126 robert deegan: the discovery of neptune
Figure 16.1: Predicted and Actual Orbits of Uranus and Neptune. Grey =
Uranus’ predicted position, light blue = Uranus’ actual position, dark blue =
Neptune’s actual position, yellow = Le Verrier’s prediction, green = Adams’
prediction [94]
Given that one object is much more massive than the other, this
attractive force will cause the less massive object to accelerate towards
the more massive object while having little effect on the latter. However,
if the less massive object is not at rest but rather has some initial velocity
vector and thus an initial angular momentum, then this force will not
cause the lesser mass to accelerate directly towards the larger mass
but rather cause it orbit about the larger object due to the necessity to
conserve the angular momentum of the system. This obviously is the
reason the planets in our solar system orbit about the Sun, due to its
relatively large mass. However, the orbits of the planets in our solar
system are not just affected by the Sun, in most cases the gravitational
field of adjacent planets are also strong enough to influence each others
orbit, as we shall see momentarily.
Shortly after Uranus was discovered in 1781 its predicted orbit was
calculated using Newton’s laws of motion and gravitation. However,
by looking at data for the position of Uranus from 1754-830(some
observations of the planet were made prior to the discovery that it
was in fact a planet) and specifically over the period from 1818-1826 it
was noticed that there were discrepancies between Uranus’ predicted
and observed orbit. Tables predicting Uranus’ orbit had been made
shortly after its discovery and soon comparison to these tables noted
discrepancies to large to be accounted for by observational error or
by the effects of Saturn and Jupiter on Uranus. As these discrepancies
grew larger and larger in the early 1800’s they became more and more
troubling, and numerous theories to account for this were postulated,
one being that there was some unknown body perturbing the orbit. This
theory appealed to two men, Urbain Le Verrier and John Couch Adams.
Both Adams and Le Verrier set out independently to investigate it.
16.4 perturbation 127
Observations had shown that not only was Uranus not at the correct
position in it’s orbit at a given time, but also that its distance from
the sun was substantially larger than it was predicted to be at some
points. This evidence to some further supported the hypothesis of the
unknown planet, and the increased radius vector clearly showed that
this mysterious planet must be outside of Uranus’ orbit. This is the
assumption Adams made as he began his investigation, but he quickly
ran into trouble. The typical way of determining a perturbing bodies
effect on an orbit was to calculate the gravitational effects of a body of
known mass and position on another planet, and subtract these values
from the observed orbit. After this had been done one could see the
"true" orbit of the planet, as it would be without the perturbing body,
and then when adding back in the effects of the perturbing planet
one gets the actual elliptical orbit followed by the planet and can thus
predict its location with extreme precision.
In this case however, the characteristics of the perturbing body were
unknown which made it impossible to calculate the "true" elliptical or-
bit, as this required knowing the perturbative effects. And these effects
in turn could only be found once one already knew the "true" orbit, so
Adams’ approach was to solve both problems simultaneously. Using
this approach Adams soon ran into another problem, that being that
the perturbation caused by a more massive planet further away from
Uranus would be indistinguishable from the effect of a less massive
planet closer to Uranus. To avoid this issue Adams simply guessed that
the average distance from the sun of the perturbing planet was twice
that of Uranus, and then to come back and adjust this assumption after
he had completed his calculations were it necessary to do so. From
this point on Adams simply needed to right out a serious of equations
relating the predicted position at each date to the actual position it was
found at. Adams wrote 21 such equations, one for every third year for
which he had observational data, and solved each of these equations
one at a time. Alone these were not enough to give the characteristics
of the perturbing planet, but each narrowed down the possibilities
further and further until after all 21 equations had been solved Adams
had a very accurate model describing the characteristics of this new
planet. Adams then calculated the effect this new planet would have on
Uranus’ orbit and found that it accounted for all of the discrepancies
between Uranus’ actual and predicted orbits. Finally, Adams calculated
the longitudinal position he expected to find Uranus at and gave this
data to the Royal Astronomer in order to confirm the existence of this
new planet there, and indeed the new planet was observed to be within
a few arcseconds of Adams’ prediction[77].
At the same time in France, Le Verrier was going through the same
processes and calculations and came to an almost identical result by
which astronomers in France independently discovered this new planet
at almost the exact same time that this was being done in England.
16.4 perturbation
2. We can apply Kepler’s 3rd Law here to find what effect this
increase in orbital distance will have on the period:
4π 2 ∗ a3
P2 = G ( M+m)
a3 = (3 × 1012 )3
= 2.7 × 1037 m3
G = 6.67 × 10 s−2 m3 kg−1
− 11
M = 1.99 × 1030 kg
m = 1.02 × 1026 kg
1.066×1039 m3
P2 = 1.327×1020 s−2 ∗m3
= 8.03 × 1018 s2
P= 2.83 × 109 s = 89.8 years
A L E X K I R I A K O P O U L O S : W H I T E D WA R F S 17
17.1 introduction
p
2mEx,y,z
k x,y,z = (17.1)
h̄
131
132 alex kiriakopoulos: white dwarfs
h̄2 k2
Enx,y,z = (17.3)
2m
where k is the magnitude of the wave vector k and each state in this
3
system occupies a volume πV for n = 1. Since the electrons behave as
identical fermions and hence are subject to the Pauli exclusion principle
only two can occupy any given state. Therefore they fill only one octant
of a sphere of k-space [37]
1 4 3 Nq π 3
( πk F ) = ( ) (17.4)
8 3 2 V
1
where k3F = (3ρπ 2 ) 3 is the radius determined by the required volume,
π3
V , the electrons must take, N is the total number of electrons, and q is
the number of free electrons each atom contributes.
Nq
Then ρ ≡ V is the free electron density. The Fermi surface here is
the boundary separating the occupied and unoccupied states in k-space.
The corresponding energy is called the Fermi energy the energy of the
highest occupied quantum state in this system of electrons. The Fermi
h̄2 2 23
energy is EF = 2m (3ρπ ) , and this is the energy for a free electron
gas.
5
h̄2 (3π 2 Nq) 3 − 2
Etotalelectron = V 3 (17.5)
10π 2 m
This energy is analogous to the internal thermal energy of ordinary
gas and exerts a pressure on the boundaries if it expands by a small
amount [37], dV.
dE = − 23 E dV
V resembles that of the work done by this electron pres-
sure,
2
2E (3π 2 ) 3 h̄2 5
P= = ρ3 (17.6)
3V 5m
Evidently the pressure depends on the free electron density. It is
when this pressure equals the gravitational pressure from the mutual
attraction of the ensemble, do white dwarf conditions occur. This is
the fundamental difference between white dwarfs and "‘normal"’ stars.
Stars usually have cores that sustain nuclear fusion of hydrogen into
helium that counteract the gravitational pressure. The white dwarf
17.3 question 133
17.3 question
h̄2 k2 V 2
dE = k dk (17.9)
2mπ 2
and
5
h̄2 (3π 2 Nq) 3 − 2
Etotal electron = V 3 (17.10)
10π 2 m
134 alex kiriakopoulos: white dwarfs
What is the integral of the above derivative. What is it? What are the
limits of integration? What does the upper limit represent?
Answer:
h̄2 V
Z kF
Etotal electron = k4 dk (17.11)
2π 2 m 0
Figure 18.1: Elements are produced at different depths within a star. This
illustrates the elements that are produced in massive stars (not to scale). Notice
the iron core.[39]
135
136 daniel rogers: supernovae and the progenitor theory
and oxygen can be further fused into neon, sodium, magnesium, sulfur
and silicon. Later reactions transform these elements into calcium, iron,
nickel, chromium, copper, etc. In a supernova event, neutron capture
reactions lead to the formation of elements heavier than iron.[53] Thus,
we see that all heavy elements are formed in the cores of stars at various
points in their lives as they burn through their thermonuclear fuel. In
general, the mass of a star can be used to determine what elements are
formed and the abundances that are produced.
E 3.8 × 1026 J
E = mc2 ⇒ m = = = 4.2 × 109 kg (18.1)
c2 (3 × 108 ms )2
This is the mass converted to energy in the sun each second. We know that
this mass is only 0.7% of the mass of the hydrogen that goes into the fusion
process.
4.2 × 109 kg
MH = = 6.0 × 1011 kg (18.2)
0.007
So we see that the sun fuses about 600 billion kg of hydrogen each second,
though about 4 billion kg are converted into energy. The remaining 596 billion
kg becomes helium.[6]
Figure 18.2: (a) The layered shells of elements in a massive star. (b) Collapse
begins when fuel is depleted. (c) As gravity takes over, the star shrinks signifi-
cantly. (d) The red area experiences enormous outward forces as hot gases pile
up on the degenerate core. (e) The gases are ejected outward at high speeds. (f)
All that remains is the degenerate neutron core.[39]
Now consider if part of this expelled cloud of isotopes (the ejecta) were
to fall into the gravitational field of a neighboring star that is in the
early stages of its formation. At this point the young star is surrounded
by what is known as a protoplanetary disk, which is the disk of gas
and dust from which planets and asteroids form. If the ejecta from
the supernova event collided and blended with this disk, it would
contribute to its chemical composition.[83]
This is one of the leading theories that explains the presence of
heavier elements in our own solar system. This is obviously impossible
to prove by experiment, but fortunately there are ways to measure the
probability that this process occurred by studying chemical properties
of our solar system.
One such property is the concentration of the stable isotope 60 Ni.
This has been detected and measured in meteorites that have been
untouched since the formation of our solar system. There is also a
correlation between the amounts of 60 Ni and another stable isotope,
56 Fe, in these meteorites.[65] The laws of chemistry predict that the
over a sphere that expands as the ejecta moves outward. Thus we will
define Ar = 4πr2 . Now let Mr60 be the total mass of 60 Fe ejected from
the supernova, and MR60 be the total mass of 60 Fe injected into our solar
system. The amount of ejecta that a solar nebula can receive is inversely
proportional to its square distance from a supernova explosion.[50]
Mr60 AR
MR60 = (18.4)
Ar
Now let M be the total mass of our solar system, and MR56 be the
amount of 56 Fe in our solar system per unit mass. Multiplying them
together gives the total amount of 56 Fe in our solar system. Then we
can define a relationship for the ratio of 60 Fe to 56 Fe injected into our
solar system.[50]
60 Fe MR60
56 Fe
= (18.5)
( MR56 M)
18.6 conclusion
18.7 problems
1. Homework Problem 1
Question: Use the result of Example Problem 1. How many times
does the proton-proton fusion reaction occur each second in the
sun? The mass of a proton is 1.6726 × 10−27 kg; hydrogen is
composed of one proton, and helium is composed of four. The
mass of a helium nucleus is 6.643 × 10−27 kg.
Solution: Here, fusion converts four hydrogen nuclei (protons)
into one helium nucleus. Four protons have a mass of 6.690 ×
10−27 . When four protons fuse to make one helium nucleus, the
amount of mass that disappears and becomes energy is as follows.
Thus, nearly 1038 fusion reactions occur in the sun each second.[6]
2. Homework Problem 2
18.7 problems 141
3. Test Problem 1
Question: Which of these elements had to be made in a supernova
explosion? (a) calcium (b) oxygen (c) uranium
Solution: (c) uranium
4. Test Problem 2
Question: Suppose there is a supernova explosion at some dis-
tance from a young solar system. If that distance is doubled, by
what factor would the radius of the solar system need to be mul-
tiplied in order to receive the same amount of ejecta? (a) 0.5 (b) 2
(c) 4
Solution: (b) 2
D AV I D PA R K E R : T H E E Q U I VA L E N C E P R I N C I P L E 19
19.1 introduction
The weak equivalence principle states that “All test particles at the
same spacetime point in a given gravitational field will undergo the
same acceleration, independent of their properties, including their
rest mass.”[92] This form of the equivalence principle is very similiar
to Einstein’s original statement on the subject, but is referred to as
“weak” because of the extent to which Einstein’s conceptualization of
the equivalence principle matured. The other two interpretations of the
equivalence principle use the weak equivalence principle as a starting
point, assuming its truth. Very few, if any, theories of gravity contradict
the weak equivalence principle.
The Einstein equivalence principle states that the result of a local non-
gravitational experiment in an inertial reference frame is independent
of the velocity of location of the experiment. This variation is basically
an extension of the Copernican idea that masses will behave exactly
the same anywhere in the universe. It grows out of the postulates of
special relativity and requires that all dimensionless physical values
remain constant everywhere in the universe.
The strong equivalence principle states that the results of any local
experiment in an inertial reference frame are independent of where and
when it is conducted. The important difference between this variation
and the first two, weaker, variations is that this is the only variation
that accounts for self-gravitating objects. That is, objects that are so
massive as to have internal gravitational interactions. Therefore, this
is an extremely important idea because of the extreme importance of
self-gravitating bodies, e.g. stars, to our understanding of the universe.
143
144 david parker: the equivalence principle
19.3 consequences
19.4 example
19.5 problems
d
F12 = − U (r ), (20.3)
dr
where F12 is the force experienced by one particle away from the other. In the
case of gravity:
d m m m m
F12 = − −G 1 2 = −G 12 2 , (20.4)
dr r r
the negative sign implying the particles move toward one another.
147
148 richard rines: the fifth interaction
“All test particles at the same spacetime point in a given gravitational field
will undergo the same acceleration, independent of their properties, including
their rest mass." [91]
More subtly, this implies that the inertial mass of an object (the propor-
tionality between the force acting on the object and its acceleration) is
exactly equal to its gravitational mass (the proportionality between the
force an object experiences in a gravitational field and the strength of
that field). This point is readily seen in observing Newton’s gravita-
tional potential (equation (20.4)).
Late in the 19th and early in the 20th century, Physicist Lorand Eötvös
set out to verify this equivalence. Between 1885 and 1909, he devised,
implemented, and refined an experiment which evidenced the equiva-
lence to a much higher degree of accuracy than had been previously
shown. Put simply, his procedure involved hanging different types of
masses in a balance along a solid rod. Torque would then be applied
to the rod from two sources: differences in gravitational force on the
two masses (a measurement of the gravitational mass of the masses),
and differences in the centrifugal force each mass experiences (a mea-
surement of the inertial mass). Only if these two masses are equivalent
would the rod remain perfectly stationary.
As with any physical procedure, experimental uncertainty and imper-
fections plagued Eötvös’s results. These were originally averaged out to
provide fairly precise results favor of weak equivalence. As subsequent
tests of the weak equivalence principle, using slightly different exper-
imental procedure, did not observe such uncertainty, the variations
in Eötvös’s results were seen as the result of imprecise experimental
process.
More recently, however, physicists such as Ephraim Fischbach began
to reexamine these irregularities in terms of a possible fifth interaction
between elementary particles.
Nq − Nq̄
B= , (20.6)
3
where Nq is the number of quarks and Nq̄ the number of antiquarks.
Specifically, Fischbach observed the fractional difference (∆κ) between
the acceleration of the object and the acceleration of gravity and found
the following linear relationship:
a B
∆κ = ∆ = α +β (20.7)
g µ
Gm1 m2 αe−r/λ
U5 (r ) = − , (20.8)
r
where α and λ are functions of the baryon numbers of m1 and m2.
5 Fischbach found that, very roughly, α = (5.65 ± 0.71) · 10−6 and β = (4.83 ± 6.44) · 10−10 .
20.5 problems 151
Figure 20.2: New experiments by the Eöt-Wash group provide results contra-
dicting Fischbach’s assumption of linearity [34]
20.5 problems
1
Answer: −4
1−7.2e−3.19·10
3. Two objects of mass 1 kg, with the same composition as Earth, are
separated by a distance of 1 mm. What force do the experience as
a result of the fifth interaction? In what direction is this force?
or 500,000 years Homo sapiens has roamed the Earth, building cities
F and creating languages. We’ve gone from stagecoaches to space
travel in the span of one human lifetime, and we’ve sent robotic scouts
to other planets. It’s difficult to imagine it all coming to an end. Yet 99
percent of all species that have ever lived have gone extinct, including
every one of our hominid ancestors. If humans survive for a long
enough time and spread throughout the galaxy, the total number of
people who will ever live might number in the trillions. It’s unlikely
that we would be among the first generation to see our descendants
colonize planets, but what are the odds that we would be the last
generation of Homo sapiens? By some estimates, the current rate of
extinctions is 10,000 times the average in the fossil record. We may be
worried about spotted owls and red squirrels, but the next statistic on
the list could be us.
Space is filled with asteroids and comets which can pose a threat to
life on Earth. Fortunately Earth’s atmosphere protects us from the
thousands of pebble-sized and smaller asteroids - only weighing a few
grams - which strike earth every day at speeds of tens of kilometers per
second1 . At these high velocities, friction with the upper atmosphere
heats the meteoroids white-hot and causes immense deceleration forces.
These small meteoroids get destroyed by the heat and deceleration, and
are seen from Earth as shooting stars. However, some of the fragments,
especially those from iron meteoroids, will reach Earth’s surface. It is
estimated that 20 tons of meteorites reach Earth’s surface each day.
So, what if an asteroid hit Earth? Not just the dust that Earth collects
as it sweeps through space, but something serious, like “Armageddon”,
or “Deep Impact”? Earth and the moon are heavily cratered from
previous impacts, the most famous of which happened on Earth at the
end of the Cretaceous period, about 65 million years ago. Scientists
hypothesize that this impact was the cause of the End-Cretaceous
(K-T) extinction, in which eighty-five percent of all species on Earth
disappeared, making it the second largest mass extinction event in
geological history2 . Dr. Richard Muller in his book “Nemesis: The
Death Star” describes the event:
153
154 douglas herbert: the science of the apocalypse
In 2028, the asteroid 1997 XF11 will come close to Earth but will miss
our planet by about two and a half lunar distances, that’s extremely
close, considering how big outer space is. If something was to change
it’s course, and it did hit Earth, what you would have is a 1.6 km (1 mile)
wide meteorite striking the planet’s surface at about 48,000 kph (30,000
mph). The energy released during the impact is related to the kinetic
energy of the asteroid before atmospheric entry begins. At typical solar
system impact speeds of 12 to 20 km/s, energy E is approximately
given as one half times the asteroid mass m times the square of the
asteroid velocity v , which can be rewritten in terms of the asteroid’s
density ρ and diameter l , assuming that the asteroid is approximately
spherical:
1 2 π 3 2
E= mv = ρl v (21.1)
2 12
A kilometer and a half wide asteroid traveling at 20 km/s has an
E roughly equal to a 1 million megaton bomb, that’s 10 million times
greater than the bomb that fell on Hiroshima. A land strike would
produce a fireball several miles wide which would briefly be as hot as
the surface of the sun, igniting anything in site, and it would blast tons
of sulfur-rich rock and dust high into the atmosphere, encircling the
globe. As the burning debris rained back down to earth, the soot and
dust would blacken the skies for months, if not years, to come. An ocean
landing would be no better, instantly vaporizing 700 cubic kilometers
(435 cubic miles) of water, and blasting a tower of steam several miles
into the atmosphere, again benighting the sky. The meteor itself would
3 The impact believed to have caused the extinction of the dinosaurs left a 300 km (186
mile) wide crater on the coast of Yucatán. The impactor had to have been at least 30 km
(19 miles) across.
21.3 errant black holes 155
likely crack Earth’s crust at the ocean floor, triggering massive global
earthquakes, sprouting volcanoes at weak spots in Earth’s crust, and
creating a tsunami as high as the ocean is deep, moving at hundreds of
kilometers an hour in every direction. Humans would likely survive,
but civilization would not.
The Kuiper belt is an asteroid zone just beyond the orbit Neptune,
and it contains roughly 100,000 ice balls that are each more than 80 and
a half kilometers (50 miles) in diameter. As Neptune’s orbit perturbs
the Kuiper belt, a steady rain of small comets are sent earthward. If one
of the big ones heads right for us, that would certainly be the end for
just about all higher forms of life on Earth.
The Milky Way galaxy is full of black holes, collapsed stars about 20 km
(12 miles) across. Just how full is hard to say, a black hole’s gravity is so
strong that anything approaching within a certain radius will no longer
be able to escape, including any light which would betray it’s presence,
this critical radius is called the “Schwarzchild radius”4 . The interior of
a sphere whose radius is the Schwarzchild radius is completely cut off
from the rest of the Universe, the only way to “see” a black hole is to
spot it’s gravitational lensing - the distortion of background light by
foreground matter, i.e., the black hole. Beyond the Schwarzchild radius,
the gravitational attraction of a black hole is indistinguishable from
that of an ordinary star of equal mass. Based on such observations,
and theoretical arguments, researchers estimate that there are about 10
million black holes in the Milky Way, including one at the core of our
galaxy5 , whose mass is as much as 2 million solar masses.
A black hole has an orbit just like any other star, so it’s not likely
that one is heading toward us. If any other star was approaching us,
we would know, with a black hole there would be little warning. A few
decades before a close approach to Earth, astronomers would probably
notice a strange perturbation in the orbits of the outer planets, as the
effects grew larger, it would be possible to make increasingly precise
estimates of the location and mass of the interloper. A black hole doesn’t
have to come very close to Earth to bring ruin, it simply needs to pass
through the solar system. Just as Neptune’s gravity disturbs the Kupier
belt, a rogue black hole could pull Earth’s orbit into an exaggerated
ellipse, either too close to the sun, or too far from the sun. It could even
eject Earth from it’s orbit and send the planet off into deep space.
of. Sixty-five million years ago, a plume of hot rock from Earth’s mantle
burst through the crust in what is now India, when continental drift
moved that area moved over a “hot spot” in the Indian Ocean. Eruptions
raged century after century, producing lava flows that exceeded 100,000
square kilometers (62 square miles) in area and 150 meters (500 feet) in
thickness. - the Laki eruption 3,000 times over. Some scientists believe
that this Indian outburst, and not an asteroid, was responsible for
the fall of the dinosaurs (the K-T extinction), since such lava flows
would have produced enormous amounts of ash, altering global climatic
conditions and changing ocean chemistry. An earlier, even larger event
in Siberia occurred at the time of the Permian mass extinction 248
million years ago, and is the most thorough extermination known to
paleontology. At that time 95 percent of all species were wiped out.
Sulfurous volcanic gases produce acid rains. Chlorine-bearing com-
pounds break down the ozone layer6 . While they are causing short-term
destruction, volcanoes also release carbon dioxide which yields long
term, greenhouse effect global warming. The last big pulse of flood
volcanism built the Columbia River plateau about 17 million years ago.
If the idea of cataclysmic volcanism sounds too unlikely, Tom Bissell
notes in a 2003 Harper’s Magazine article (A Comet’s Tale) that:
...73,500 years ago what is known as a volcanic supere-
ruption occurred in modern-day Sumatra. The resultant
volcanic winter blocked photosynthesis for years and very
nearly wiped out the human race. DNA studies have sug-
gested that the sum total of human characteristics can be
traced back to a few thousand survivors of this catastrophe.
More properly known as coronal mass ejections, solar flares are the
outbursts caused by enormous magnetic storms on the sun, and they
bombard Earth with high speed subatomic particles. A typical solar
flare releases the equivalent energy of a billion hydrogen bombs, and
ejects a hundred billion tons of high energy particles into space. Earth’s
magnetic field and atmosphere negate the potentially lethal effects of
ordinary flares, knocking the particles back into space, and steering
some of the particles over the poles, creating Earth’s auroras. But while
examining old astronomical records, Yale University’s Bradley Schaefer
found evidence that some sunlike stars can brighten briefly by up
to a factor of 20. Schaefer believes that these increases are caused by
superflares millions of times more powerful than those commonly
experienced by Earth. Scientists don’t know why superflares happen at
all, or whether our sun could exhibit milder but still disruptive behavior.
A superflare on the sun only a hundred times larger than typical would
overwhelm Earth’s magnetosphere and begin disintegrating the ozone
layer (see footnote 6). Such a burst would certainly kill anything basking
in its glow, and according to Bruce Tsurutani of NASA’s Jet Propulson
Labatory, “it would leave no trace in history.”
While too much solar activity could be deadly, too little of it is
problematic as well. Sallie Baliunas of the Harvard-Smithsonian Center
for Astrophysics says that many sunlike stars pass through extended
6 Without the ozone layer, ultraviolet rays from the sun would reach the surface the earth
at nearly full force, causing skin cancer, and killing off the photosynthetic plankton in the
ocean that provide oxygen to the atmosphere and support the bottom of the food chain.
21.6 viral epidemic 157
Humans are here today because our particular line never frac-
tured - never once at any of the billion points that could have
erased us from history.
158 douglas herbert: the science of the apocalypse
ife, as far as we know, exists only on our own planet Earth. There
L is no evidence of intelligence (or life of any form) in our solar
system, galaxy, or beyond. But are we really alone? The idea that our
small planet possesses the only biological beings in the universe seems
egotistical and, in light of our recent advances in astronomy and the
sciences, unlikely. Many efforts are currently underway in the search for
extraterrestrial intelligence, with the hope of finding it still undamped
by the lack of success. If intelligent life does indeed exist elsewhere, it
remains to be seen whether it will dwell in a region close enough to
make contact with, and even whether or not it will want to be found.
NC = RS f P n f L f I f C L, (22.1)
where
159
160 amanda lund: extraterrestrial intelligence
As the only term in this equation known with any certainty is the rate
of star formation RS (about 2 or 3 per year), the projected values of NC
vary greatly [75]. The fraction of stars with solar systems ( f P ) has been
estimated between 20% and 50% [55], but the remaining terms depend
greatly on the optimism of the person assigning them. For instance,
the average lifetime of an intelligent civilization might be anywhere
between 100 and 1010 years. A belief in the tendency of civilizations to
self-destruct would set L at the lower limit, while the opposing view
that societies can overcome this inclination to eradicate themselves
would let L be as large as the lifetime of their star [75].
As the only known example of f L , f I , and f C is the earth, there is
really no way to obtain an accurate estimate of these probabilities [57].
Based on the time it took life to evolve on earth, a rough estimate of
f L > 0.13 could be assigned. Drake has estimated both f I , and f C to be
0.01, though there is not much basis for these assumptions.
The current quest for intelligent life mainly involve filtering through
multitudes of electromagnetic emissions in an attempt to "eavesdrop"
on possible sentient societies [82]. Since World War II, radio astronomy
has advanced significantly, and most of the searches are being done in
the radio and microwave regions of the spectrum. SETI (the Search for
ExtraTerrestrial Intelligence), the main organization conducting these
searches, has been using large radio telescopes in an attempt to pick
up alien transmissions. Their receivers are intended to identify narrow-
band broadcasts, which are easy to discern even at large distances and
can only be created by transmitters [84].
A particularly important signal for astronomy and SETI is the hy-
22.4 will we find it? and when? 161
Solution:
First we must convert the energy in eV to joules, using the conversion factor
Using both this value of the change in energy in joules and Planck’s constant,
we can determine the frequency of the photon emitted using Planck’s relation.
E = hν (22.5)
∆E 9.44 × 10−25
J
ν= = = 1420 MHz (22.6)
h 6.626 × 10−34 J s
Now, by plugging this frequency into the equation relating frequency and
wavelength (where c denotes the speed of light in a vacuum), we can find the
corresponding wavelength.
c 3 × 108 m/s
λ= = = 0.21 m = 21cm (22.7)
ν 1420 MHz
Over the past 50 years, great advances in technology and in our un-
derstanding of physics and astronomy have transformed the search for
extraterrestrial intelligence from science fiction and Hollywood horror
films into an actual science. Despite this, we may still be far from find-
ing other life in the universe. If it does indeed exist, it may be nearly
162 amanda lund: extraterrestrial intelligence
Figure 22.2: SETI uses radio telescopes to try to pick up signals from extrater-
restrial life [71].
impossible to locate life among the billions of stars and galaxies and
light years of space that surround us, especially if any intelligent life out
there does not want to be found. The search has been likened to trying
to find not a needle but a single atom of a needle in a haystack [74].
Yet there are still optimists who believe we might not be so far off
after all. Notable astronomers including Carl Sagaan and Frank Drake,
who are searching through this cosmic haystack themselves, have pre-
dicted that extraterrestrial intelligence will be found between 2015 and
2026 [74]. SETI even offers a program called SETI@home, where anyone
with a computer can donate disk space to help analyze radio telescope
data and speed up the search. It seems reasonable to believe we are not
alone in the universe; the real question, then, must be whether or not
we will see evidence of other intelligence–and if so, when.
22.5 problems
Solution: Using
NC = RS f P n f L f I f C L, (22.8)
the most optimistic value would use the upper limits in the ranges, and
the result would be
3 × 108 m/s
ν= = 4 × 1014 s−1 , (22.12)
7.5 × 10−7 m
and the lower frequency is
3 × 108 m/s
ν= = 3 × 1013 s−1 , (22.13)
1 × 10−5 m
22.7 summary
So far, the only life we know of in the universe exists on Earth. However,
given the vastness of the universe and the huge number of galaxies,
stars, and planets within it, it is unlikely that we are alone. With our
current technology there is no way to tell how much intelligent life
exists outside our planet or where it is located; the Drake Equation is
one way to estimate the number of intelligent civilizations, but it is
based mainly on guesswork. The SETI institute has set out to find life by
attempting to eavesdrop on radio signals from other civilizations–one
important wavelength in this search is the 21-centimeter line, a rare
transition between two hyperfine states of hydrogen. There is no way
to know when or if we will find anything, though some optimistic
astronomers predict as soon as 2015.
BIBLIOGRAPHY A
[1] The Eöt-Wash Group, Washington University, 2008.
http://www.npl.washington.edu/eotwash/. (Cited on
page 148.)
[7] Ben Best. Lessons for cryonics from metallurgy and ceram-
ics. benbest.com/cryonics/lessons.html. (Cited on pages 34
and 35.)
[16] Kenneth Chang. The nature of glass remains anything but clear,
July 2008. New York Times. (Cited on pages 33 and 34.)
165
166 bibliography
[23] Marco Dorigo. Ant colonies for the traveling salesman problem.
Technical report, Université Libre de Bruxelles, 1996. (Cited on
pages 111 and 113.)
[24] Marco Dorigo. Ant algorithms for discrete optimization. Technical
report, Université Libre de Bruxelles, 1996. (Cited on pages 111
and 112.)
[25] Belle Dumé. Magnetic recording has a speed limit. http:
//physicsworld.com/cws/article/news/19401, Apr 21, 2004.
(Cited on page 87.)
[26] Albert Einstein. Relativity: The special and general theory. Henry
Holt and Company, New York, 1920. Translated by Robert W.
Lawson. (Cited on page 143.)
[27] SR Elliot. Physics of Amorphous Materials. Longman, 1984. (Cited
on pages 33, 34, and 35.)
[28] B. N. J. Persson et. al. On the origin of amonton’s friction law.
Journal of Physics: Condensed Matter, 20, 2008. (Cited on page 57.)
[29] H. Olsson et. al. Friction models and friction compensation.
European Journal of Control, 4(9), 1998. (Cited on pages 56 and 57.)
[30] Extrasolar Planets. Extrasolar planets. http://www.astro.keele.
ac.uk/workx/extra-solar-planets/extrasolar.html. Image of
an artist’s representation of an extrasolar planet taken from this
site. (Cited on page 160.)
[31] Adalbert Feltz. Amorphous Inorganic Materials and Glasses. VCH,
1993. (Cited on page 33.)
[32] Richard Feynman, Robert Leighton, and Matthew Sands. The
Feynman Lectures on Physics. Addison-Wesley, San Franscisco,
1964. (Cited on page 19.)
[33] Ephraim Fischbach. Reanalysis of the Eötvös Experiment. 1986.
(Cited on page 149.)
[34] Allan Franklin. The Rise and Fall of the Fifth Force. American Insti-
tute of Physics, New York, New York, 1993. (Cited on pages 148,
149, 150, and 151.)
bibliography 167
[35] Carolyn Gordon, David L. Webb, and Scott Wolpert. One cannot
hear the shape of a drum. Bulletin of the American Mathematical
Society, 27:134–138, 1992. (Cited on page 29.)
[36] David J. Griffiths. Introduction to Electrodynamics. Prentice Hall,
Upper Saddle River, New Jersey, 3rd edition, 1999. (Cited on
pages 3, 5, 6, and 8.)
[37] David J. Griffiths. Introduction to Quantum Mechanics. Benjamin
Cummings, San Francisco, second edition, 2004. (Cited on
pages 61, 62, 64, and 132.)
[38] John Grue and Karsten Trulsen. Waves in Geophysical Fluids.
SpringerWienNewYork, Italy, 2006. (Cited on page 47.)
[39] R. J. Hall. Interior structure of star prior to core collapse.
http://www.newwest.net/topic/article/cosmicfireworks, De-
cember 2006. (Cited on pages 135 and 137.)
[40] R.V.L. Hartley. Transmission of information. Bell System Technical
Journal, July:535–563, 1928. (Cited on page 98.)
[41] Sverre Haver. Freak wave event at draupner jacket. Statoil, 2005.
(Cited on page 48.)
[42] Franz Himpsel. Nanoscale memory. http://uw.physics.wisc.
edu/~himpsel/memory.html, 2008. (Cited on page 88.)
[57] Roy Mash. Big numbers and induction in the case for extraterres-
trial intelligence. Philosophy of Science, 60:204–222, 1993. (Cited
on page 160.)
[58] S.R. Massel. Ocean Surface Waves; their Physics and Prediction.
World Scientific Publ., New Jersey, 1996. (Cited on page 49.)
[60] Kurt Nassau. The Physics and Chemistry of Color. John Wiley and
Sons, Inc., New York, second edition, 2001. (Cited on page 9.)
[74] Seth Shostak. When will we find the extraterrestrials? (And What
Will We Do When We Find Them?). lecture at the University of
Massachusetts Amherst, October 2008. (Cited on page 162.)
[76] Richard Slansky, Stuart Raby, Terry Goldman, and Gerry Garvey.
The oscillating neutrino: An introduction to neutrino masses and
mixings. Los Alamos Science, 25:28–63, 1997. (Cited on pages 65,
66, 72, 73, and 74.)
[80] R.J. Tayler. The Stars: their structure and evolution. Cambridge,
1994. (Cited on page 133.)
[97] Lei Xu, Wendy W. Zhang, and Sidney R. Nagel. Drop splashing
on a dry smooth surface. Phys. Rev. Lett., 94, 2005. (Cited on
page 39.)
bibliography 171
[98] Lei Xu, Loreto Barcos, and Sidney R. Nagel. Splashing of liquids:
Interplay of surface roughness with surrounding gas. Phys. Rev.,
76, 2007. (Cited on page 39.)
Andersen [2], 17, 165 Limongi and Chieffi [51], 138, 139,
Arndt [3], 98, 165 167
BBC [4], 47, 53, 165 Lloyd [52], 102, 167
Bender and Orszag [5], 29, 165 Lochner and Gibb [53], 135–137,
Bennet et al. [6], 135–137, 140, 165 168
Best [7], 34, 35, 165 Lyons [54], 79, 80, 83, 168
Bleaney [8], 4, 165 Marais and Walter [55], 159, 160,
Bloom [9], 4, 7, 165 168
Blum [10], 109, 165 Margolus and Levitin [56], 88, 168
Brawer [11], 34, 165 Mash [57], 160, 168
Brown [12], 47, 165 Massel [58], 49, 168
Bryant [13], 47, 165 Mermin [59], xi, 168
Cadonati [14], 64, 66, 67, 72, 74, 165 Nassau [60], 9, 168
Cadonati [15], 161, 165 Nave [61], 135, 168
Chang [16], 33, 34, 165 Nave [62], 4, 168
Colloid [17], 36, 165 News [63], 89, 168
Commons [18], 128, 166 Palme and Jones [65], 138, 168
Cowles [19], 128, 166 Peters et al. [66], 107, 108, 168
Cross [20], 56, 166 Pinet [67], 47, 168
Danby [21], 128, 166 Research [68], 91, 93, 168
Dorigo [22], 107, 109, 166 Royall et al. [69], 36, 37, 169
Dorigo [23], 111, 113, 166 SETI [71], 162, 169
Dorigo [24], 111, 112, 166 Sears and Zemansky [70], 10, 169
Dumé [25], 87, 166 Shannon [72], 98, 169
Einstein [26], 143, 166 Shapiro and Teukolsky [73], 131,
Elliot [27], 33–35, 166 169
Extrasolar Planets [30], 160, 166 Shostak [74], 162, 169
Feltz [31], 33, 166 Shu [75], 160, 169
Feynman et al. [32], 19, 166 Slansky et al. [76], 65, 66, 72–74,
Fischbach [33], 149, 166 169
Franklin [34], 148–151, 166 Standage [77], 127, 169
Gordon et al. [35], 29, 166 Svistunov [78], 64, 169
Griffiths [36], 3, 5, 6, 8, 167 Tachibana et al. [79], 139, 169
Griffiths [37], 61, 62, 64, 132, 167 Tayler [80], 133, 169
Grue and Trulsen [38], 47, 167 Than [81], 138, 169
Hall [39], 135, 137, 167 Thomsen [82], 160, 169
Hartley [40], 98, 167 Tobin et al. [83], 138, 169
Haver [41], 48, 167 Townes [84], 160, 163, 170
Himpsel [42], 88, 167 Tsymbal [85], 94, 170
Iizuka [43], 13, 167 Turner [86], 107, 108, 170
Ilardi [44], 13, 167 University [87], 144, 170
Kac [45], 27, 29, 167 Veltman [88], 62, 63, 65, 170
Kamen and Heck [46], 81, 84, 167 Wallace [89], 12, 170
Kasper and Feller [47], 19–21, 23, Waugh [90], 39, 170
167 Wesson [91], 148, 170
Kharif and Pelinovsky [48], 52, 53, Wesson [92], 143, 170
167 Whitaker [93], 10, 11, 170
Lavrenov [49], 50, 51, 167 William Sheehan and Waff [94], 126,
Leshin et al. [50], 137–140, 167 170
173
174 INDEX