20–1Operations and operators
All the things we have done so far in quantum mechanics could be
handled with ordinary algebra, although we did from time to time show
you some special ways of writing quantum-mechanical quantities and
equations. We would like now to talk some more about some interesting
and useful mathematical ways of describing quantum-mechanical
things. There are many ways of approaching the subject of quantum
mechanics, and most books use a different approach from the one we
have taken. As you go on to read other books you might not see right
away the connections of what you will find in them to what we have
been doing. Although we will also be able to get a few useful results,
the main purpose of this chapter is to tell you about some of the
different ways of writing the same physics. Knowing them you should be
able to understand better what other people are saying. When people
were first working out classical mechanics they always wrote all the
equations in terms of x-, y-, and z-components. Then someone
came along and pointed out that all of the writing could be made much
simpler by inventing the vector notation. It’s true that when you come
down to figuring something out you often have to convert the vectors
back to their components. But it’s generally much easier to see what’s
going on when you work with vectors and also easier to do many of the
calculations. In quantum mechanics we were able to write many things
in a simpler way by using the idea of the “state vector.” The state
vector |ψ⟩ has, of course, nothing to do with geometric
vectors in three dimensions but is an abstract symbol that
stands for a physical state, identified by the “label,” or
“name,” ψ. The idea is useful because the laws of quantum
mechanics can be written as algebraic equations in terms of these
symbols. For instance, our fundamental law that any state can be made
up from a linear combination of base states is written as
|ψ⟩=∑iCi|i⟩,(20.1)
where the
Ci are a set of ordinary (complex) numbers—the
amplitudes
Ci=⟨i|ψ⟩—while
|1⟩,
|2⟩,
|3⟩, and so on, stand for the base states in some base, or
representation.
If you take some physical state and do something to it—like rotating
it, or like waiting for the time Δt—you get a different
state. We say, “performing an operation on a state produces a new
state.” We can express the same idea by an equation:
|ϕ⟩=A^|ψ⟩.(20.2)
An operation on a state produces another state. The
operator A^ stands for some particular operation. When this
operation is performed on any state, say
|ψ⟩, it produces some
other state
|ϕ⟩.
What does Eq. (20.2) mean? We define it this
way. If you multiply the equation by ⟨i| and expand |ψ⟩
according to Eq. (20.1), you get
⟨i|ϕ⟩=∑j⟨i|A^|j⟩⟨j|ψ⟩.(20.3)
(The states
|j⟩ are from the same set as
|i⟩.) This is now
just an algebraic equation. The numbers
⟨i|ϕ⟩ give the
amount of each base state you will find in
|ϕ⟩, and it is
given in terms of a linear superposition of the
amplitudes
⟨j|ψ⟩ that you find
|ψ⟩ in each base state. The
numbers
⟨i|A^|j⟩ are just the coefficients which tell
how much of
⟨j|ψ⟩ goes into each sum. The operator
A^
is described numerically by the set of numbers, or “matrix,”
Aij≡⟨i|A^|j⟩.(20.4)
So Eq. (20.2) is a high-class way of writing
Eq. (20.3). Actually it is a little more than that;
something more is implied. In Eq. (20.2) we do not make
any reference to a set of base states. Equation (20.3) is
an image of Eq. (20.2) in terms of some set of base
states. But, as you know, you may use any set you wish. And this idea is
implied in Eq. (20.2). The operator way of writing avoids
making any particular choice. Of course, when you want to get definite
you have to choose some set. When you make your choice, you use
Eq. (20.3). So the operator
equation (20.2) is a more abstract way of writing the
algebraic equation (20.3). It’s similar to the
difference between writing
c=a×b
instead of
cxcycz=aybz−azby,=azbx−axbz,=axby−aybx.
The first way is much handier. When you want
results, however,
you will eventually have to give the components with respect to some
set of axes. Similarly, if you want to be able to say what you really
mean by
A^, you will have to be ready to give the matrix
Aij
in terms of
some set of base states. So long as you have in
mind some set
|i⟩, Eq. (
20.2) means just the same
as Eq. (
20.3). (You should remember also that once you
know a matrix for one particular set of base states you can always
calculate the corresponding matrix that goes with any other base. You
can transform the matrix from one “representation” to another.)
The operator equation in (20.2) also allows a new way of
thinking. If we imagine some operator A^, we can use it with any
state |ψ⟩ to create a new state A^|ψ⟩. Sometimes
a “state” we get this way may be very peculiar—it may not
represent any physical situation we are likely to encounter in
nature. (For instance, we may get a state that is not normalized to
represent one electron.) In other words, we may at times get
“states” that are mathematically artificial. Such artificial
“states” may still be useful, perhaps as the mid-point of some
calculation.
We have already shown you many examples of quantum-mechanical
operators. We have had the rotation operator R^y(θ) which
takes a state |ψ⟩ and produces a new state, which is the old
state as seen in a rotated coordinate system. We have had the parity
(or inversion) operator P^, which makes a new state by reversing
all coordinates. We have had the operators σ^x, σ^y,
and σ^z for spin one-half particles.
The operator J^z was defined in Chapter 17 in terms
of the rotation operator for a small angle ϵ.
R^z(ϵ)=1+iℏϵJ^z.(20.5)
This just means, of course, that
R^z(ϵ)|ψ⟩=|ψ⟩+iℏϵJ^z|ψ⟩.(20.6)
In this example,
J^z|ψ⟩ is
ℏ/iϵ times the
state you get if you rotate
|ψ⟩ by the small angle
ϵ
and then subtract the original state. It represents a “state” which
is the
difference of two states.
One more example. We had an operator p^x—called the momentum
operator (x-component) defined in an equation
like (20.6). If D^x(L) is the operator which displaces
a state along x by the distance L, then p^x is defined by
D^x(δ)=1+iℏδp^x,(20.7)
where
δ is a small displacement. Displacing the
state
|ψ⟩ along
x by a small distance
δ gives a new
state
|ψ′⟩. We are saying that this new state is the old state
plus a small new piece
iℏδp^x|ψ⟩.
The operators we are talking about work on a state vector
like |ψ⟩, which is an abstract description of a physical
situation. They are quite different from algebraic
operators which work on mathematical functions. For instance, d/dx is an
“operator” that works on f(x) by changing it to a new
function f′(x)=df/dx. Another example is the algebraic
operator ∇2. You can see why the same word is used in both
cases, but you should keep in mind that the two kinds of operators are
different. A quantum-mechanical operator A^ does not work on
an algebraic function, but on a state vector like |ψ⟩. Both
kinds of operators are used in quantum mechanics and often in similar
kinds of equations, as you will see a little later. When you are first
learning the subject it is well to keep the distinction always in mind.
Later on, when you are more familiar with the subject, you will find
that it is less important to keep any sharp distinction between the two
kinds of operators. You will, indeed, find that most books generally use
the same notation for both!
We’ll go on now and look at some useful things you can do with
operators. But first, one special remark. Suppose we have an
operator A^ whose matrix in some base
is Aij≡⟨i|A^|j⟩. The amplitude that the
state A^|ψ⟩ is also in some other state |ϕ⟩
is ⟨ϕ|A^|ψ⟩. Is there some meaning to the complex
conjugate of this amplitude? You should be able to show that
⟨ϕ|A^|ψ⟩∗=⟨ψ|A^†|ϕ⟩,(20.8)
where
A^† (read “A dagger”) is an operator whose matrix
elements are
A†ij=(Aji)∗.(20.9)
To get the
i,j element of
A† you go to the
j,i element
of
A (the indexes are reversed) and take its complex conjugate. The
amplitude that the state
A^†|ϕ⟩ is in
|ψ⟩ is
the complex conjugate of the amplitude that
A^|ψ⟩ is
in
|ϕ⟩. The operator
A^† is called the
“Hermitian adjoint” of
A^.
Many important operators of quantum mechanics have the special property
that when you take the Hermitian adjoint, you get the same operator
back. If
B^ is such an operator, then
B^†=B^,
and it is called a “self-adjoint” or “Hermitian,” operator.
20–2Average energies
So far we have reminded you mainly of what you already know. Now we
would like to discuss a new question. How would you find the
average energy of a system—say, an atom? If an atom is in a
particular state of definite energy and you measure the energy, you
will find a certain energy E. If you keep repeating the measurement
on each one of a whole series of atoms which are all selected to be in
the same state, all the measurements will give E, and the
“average” of your measurements will, of course, be just E.
Now, however, what happens if you make the measurement on some
state |ψ⟩ which is not a stationary state? Since the system
does not have a definite energy, one measurement would give one
energy, the same measurement on another atom in the same state would
give a different energy, and so on. What would you get for the average
of a whole series of energy measurements?
We can answer the question by projecting the state |ψ⟩ onto
the set of states of definite energy. To remind you that this is a
special base set, we’ll call the states |ηi⟩. Each of the
states |ηi⟩ has a definite energy Ei. In this
representation,
|ψ⟩=∑iCi|ηi⟩.(20.10)
When you make an energy measurement and get some number
Ei, you
have found that the system was in the state
ηi. But you may get
a different number for each measurement. Sometimes you will get
E1,
sometimes
E2, sometimes
E3, and so on. The
probability
that you observe the energy
E1 is just the probability of finding
the system in the state
|η1⟩, which is, of course, just the
absolute square of the amplitude
C1=⟨η1|ψ⟩. The
probability of finding each of the possible energies
Ei is
Pi=|Ci|2.(20.11)
How are these probabilities related to the mean value of a whole
sequence of energy measurements? Let’s imagine that we get a series of
measurements like this: E1, E7, E11, E9, E1,
E10, E7, E2, E3, E9, E6, E4, and so on. We
continue for, say, a thousand measurements. When we are finished we
add all the energies and divide by one thousand. That’s what we mean
by the average. There’s also a short-cut to adding all the
numbers. You can count up how many times you get E1, say that
is N1, and then count up the number of times you get E2, call
that N2, and so on. The sum of all the energies is certainly just
N1E1+N2E2+N3E3+⋯=∑iNiEi.
The average energy is this sum divided by the total number of
measurements which is just the sum of all the
Ni's, which we can
call
N;
Eav=∑iNiEiN.(20.12)
We are almost there. What we mean by the probability of
something happening is just the number of times we expect it to happen
divided by the total number of tries. The ratio Ni/N should—for
large N—be very near to Pi, the probability of finding the
state |ηi⟩, although it will not be exactly Pi because of
the statistical fluctuations. Let’s write the predicted (or
“expected”) average energy as ⟨E⟩av; then we can say that
⟨E⟩av=∑iPiEi.(20.13)
The same arguments apply for any measurement. The average value of a
measured quantity
A should be equal to
⟨A⟩av=∑iPiAi,
where
Ai are the various possible values of the observed quantity,
and
Pi is the probability of getting that value.
Let’s go back to our quantum-mechanical state |ψ⟩. Its
average energy is
⟨E⟩av=∑i|Ci|2Ei=∑iC∗iCiEi.(20.14)
Now watch this trickery! First, we write the sum as
∑i⟨ψ|ηi⟩Ei⟨ηi|ψ⟩.(20.15)
Next we treat the left-hand
⟨ψ| as a common “factor.” We
can take this factor out of the sum, and write it as
⟨ψ|{∑i|ηi⟩Ei⟨ηi|ψ⟩}.
This expression has the form
⟨ψ|ϕ⟩,
where
|ϕ⟩ is some “cooked-up” state defined by
|ϕ⟩=∑i|ηi⟩Ei⟨ηi|ψ⟩.(20.16)
It is, in other words, the state you get if you take each base
state
|ηi⟩ in the amount
Ei⟨ηi|ψ⟩.
Now remember what we mean by the states |ηi⟩. They are
supposed to be the stationary states—by which we mean that for each
one,
H^|ηi⟩=Ei|ηi⟩.
Since
Ei is just a number, the right-hand side is the same
as
|ηi⟩Ei, and the sum in Eq. (
20.16) is the
same as
∑iH^|ηi⟩⟨ηi|ψ⟩.
Now
i appears only in the famous combination that contracts to
unity, so
∑iH^|ηi⟩⟨ηi|ψ⟩=H^∑i|ηi⟩⟨ηi|ψ⟩=H^|ψ⟩.
Magic! Equation (
20.16) is the same as
|ϕ⟩=H^|ψ⟩.(20.17)
The average energy of the state
|ψ⟩ can be written very
prettily as
⟨E⟩av=⟨ψ|H^|ψ⟩.(20.18)
To get the average energy you operate on
|ψ⟩ with
H^, and
then multiply by
⟨ψ|. A simple result.
Our new formula for the average energy is not only pretty. It is also
useful, because now we don’t need to say anything about any particular
set of base states. We don’t even have to know all of the possible
energy levels. When we go to calculate, we’ll need to describe our
state in terms of some set of base states, but if we know the
Hamiltonian matrix Hij for that set we can get the average
energy. Equation (20.18) says that for any set of
base states |i⟩, the average energy can be calculated from
⟨E⟩av=∑ij⟨ψ|i⟩⟨i|H^|j⟩⟨j|ψ⟩,(20.19)
where the amplitudes
⟨i|H^|j⟩ are just the elements of
the matrix
Hij.
Let’s check this result for the special case that the states |i⟩
are the definite energy states. For them,
H^|j⟩=Ej|j⟩,
so ⟨i|H^|j⟩=Ejδij and
⟨E⟩av=∑ij⟨ψ|i⟩Ejδij⟨j|ψ⟩=∑iEi⟨ψ|i⟩⟨i|ψ⟩,
which is right.
Equation (20.19) can, incidentally, be extended to other
physical measurements which you can express as an operator. For
instance, L^z is the operator of the z-component of the angular
momentum L. The average of the z-component for the
state |ψ⟩ is
⟨Lz⟩av=⟨ψ|L^z|ψ⟩.
One way to prove it is to think of some situation in which the energy
is proportional to the angular momentum. Then all the arguments go
through in the same way.
In summary, if a physical observable A is related to a suitable
quantum-mechanical operator A^, the average value of A for the
state |ψ⟩ is given by
⟨A⟩av=⟨ψ|A^|ψ⟩.(20.20)
By this we mean that
⟨A⟩av=⟨ψ|ϕ⟩,(20.21)
with
|ϕ⟩=A^|ψ⟩.(20.22)
20–3The average energy of an atom
Suppose we want the average energy of an atom in a state described by
a wave function ψ(r); How do we find it? Let’s first think of
a one-dimensional situation with a state |ψ⟩ defined by the
amplitude ⟨x|ψ⟩=ψ(x). We are asking for the special
case of Eq. (20.19) applied to the coordinate
representation. Following our usual procedure, we replace the states
|i⟩ and |j⟩ by |x⟩ and |x′⟩, and change the
sums to integrals. We get
⟨E⟩av=∬⟨ψ|x⟩⟨x|H^|x′⟩⟨x′|ψ⟩dxdx′.(20.23)
This integral can, if we wish, be written in the following way:
∫⟨ψ|x⟩⟨x|ϕ⟩dx,(20.24)
with
⟨x|ϕ⟩=∫⟨x|H^|x′⟩⟨x′|ψ⟩dx′.(20.25)
The integral over
x′ in (
20.25) is the same one we had
in Chapter
16—see Eq. (
16.50) and
Eq. (
16.52)—and is equal to
−ℏ22md2dx2ψ(x)+V(x)ψ(x).
We can therefore write
⟨x|ϕ⟩={−ℏ22md2dx2+V(x)}ψ(x).(20.26)
Remember that ⟨ψ|x⟩= ⟨x|ψ⟩∗= ψ∗(x);
using this equality, the average energy in Eq. (20.23)
can be written as
⟨E⟩av=∫ψ∗(x){−ℏ22md2dx2+V(x)}ψ(x)dx.(20.27)
Given a wave function
ψ(x), you can get the average energy by
doing this integral. You can begin to see how we can go back and forth
from the state-vector ideas to the wave-function ideas.
The quantity in the braces of Eq. (20.27) is an
algebraic operator. We will write it as H^
H^=−ℏ22md2dx2+V(x).
With this notation Eq. (
20.23) becomes
⟨E⟩av=∫ψ∗(x)H^ψ(x)dx.(20.28)
The algebraic operator H^ defined here is, of course, not
identical to the quantum-mechanical operator H^. The new operator
works on a function of position ψ(x)=⟨x|ψ⟩ to give a
new function of x, ϕ(x)=⟨x|ϕ⟩; while H^ operates
on a state vector |ψ⟩ to give another state
vector |ϕ⟩, without implying the coordinate representation or any
particular representation at all. Nor is H^ strictly the same
as H^ even in the coordinate representation. If we choose to work
in the coordinate representation, we would interpret H^ in terms
of a matrix ⟨x|H^|x′⟩ which depends somehow on the two
“indices” x and x′; that is, we expect—according to
Eq. (20.25)—that ⟨x|ϕ⟩ is related to all the
amplitudes ⟨x|ψ⟩ by an integration. On the other hand, we
find that H^ is a differential operator. We have already worked
out in Section 16–5 the connection
between ⟨x|H^|x′⟩ and the algebraic operator H^.
We should make one qualification on our results. We have been assuming
that the amplitude ψ(x)=⟨x|ψ⟩ is normalized. By this
we mean that the scale has been chosen so that
∫|ψ(x)|2dx=1;
so the probability of finding the electron
somewhere is
unity. If you should choose to work with a
ψ(x) which is not
normalized you should write
⟨E⟩av=∫ψ∗(x)H^ψ(x)dx∫ψ∗(x)ψ(x)dx.(20.29)
It’s the same thing.
Notice the similarity in form between Eq. (20.28) and
Eq. (20.18). These two ways of writing the same result
appear often when you work with the x-representation. You can go
from the first form to the second with any A^ which is a
local operator, where a local operator is one which in the
integral
∫⟨x|A^|x′⟩⟨x′|ψ⟩dx′
can be written as
A^ψ(x), where
A^ is a differential
algebraic operator. There are, however, operators for which this is
not true. For them you must work with the basic equations in
(
20.21) and (
20.22).
You can easily extend the derivation to three dimensions. The result
is that
⟨E⟩av=∫ψ∗(r)H^ψ(r)dV,(20.30)
with
H^=−ℏ22m∇2+V(r),(20.31)
and with the understanding that
∫|ψ|2dV=1.(20.32)
The same equations can be extended to systems with several electrons
in a fairly obvious way, but we won’t bother to write down the
results.
With Eq. (20.30) we can calculate the average
energy of an atomic state even without knowing its energy levels. All we
need is the wave function. It’s an important law. We’ll tell you about
one interesting application. Suppose you want to know the ground-state
energy of some system—say the helium atom, but it’s too hard to solve
Schrödinger’s equation for the wave function, because there are too many
variables. Suppose, however, that you take a guess at the wave
function—pick any function you like—and calculate the average
energy. That is, you use Eq. (20.29)—generalized to three
dimensions-to find what the average energy would be if the atom were
really in the state described by this wave function. This energy will
certainly be higher than the ground-state energy which is the
lowest possible energy the atom can have.
Now pick another function and
calculate its average energy. If it is lower than your first choice
you are getting closer to the true ground-state energy. If you keep on
trying all sorts of artificial states you will be able to get lower
and lower energies, which come closer and closer to the ground-state
energy. If you are clever, you will try some functions which have a
few adjustable parameters. When you calculate the energy it will be
expressed in terms of these parameters. By varying the parameters to
give the lowest possible energy, you are trying out a whole class of
functions at once. Eventually you will find that it is harder and
harder to get lower energies and you will begin to be convinced that
you are fairly close to the lowest possible energy. The helium atom
has been solved in just this way—not by solving a differential
equation, but by making up a special function with a lot of adjustable
parameters which are eventually chosen to give the lowest possible
value for the average energy.
20–4The position operator
What is the average value of the position of an electron in an atom?
For any particular state |ψ⟩ what is the average value of the
coordinate x? We’ll work in one dimension and let you extend the
ideas to three dimensions or to systems with more than one particle:
We have a state described by ψ(x), and we keep measuring x over
and over again. What is the average? It is
∫xP(x)dx,
where
P(x)dx is the probability of finding the electron in a
little element
dx at
x. Suppose the probability density
P(x)
varies with
x as shown in Fig.
20–1.
The electron is most likely
to be found near the peak of the curve. The average value of
x is
also somewhere near the peak. It is, in fact, just the center of
gravity of the area under the curve.
We have seen earlier that P(x) is
just |ψ(x)|2=ψ∗(x)ψ(x), so we can write the average
of x as
⟨x⟩av=∫ψ∗(x)xψ(x)dx.(20.33)
Our equation for ⟨x⟩av has the same form as Eq. (20.28).
For the average energy, the energy operator H^ appears between
the two ψ’s, for the average position there is just x. (If you
wish you can consider x to be the algebraic operator “multiply
by x.”) We can carry the parallelism still further, expressing the
average position in a form which corresponds to
Eq. (20.18). Suppose we just write
⟨x⟩av=⟨ψ|α⟩(20.34)
with
|α⟩=x^|ψ⟩,(20.35)
and then see if we can find the operator
x^ which generates the
state
|α⟩, which will make Eq. (
20.34) agree
with Eq. (
20.33). That is, we must find
a
|α⟩, so that
⟨ψ|α⟩=⟨x⟩av=∫⟨ψ|x⟩x⟨x|ψ⟩dx.(20.36)
First, let’s expand
⟨ψ|α⟩ in the
x-representation. It is
⟨ψ|α⟩=∫⟨ψ|x⟩⟨x|α⟩dx.(20.37)
Now compare the integrals in the last two equations. You see that in
the
x-representation
⟨x|α⟩=x⟨x|ψ⟩.(20.38)
Operating on
|ψ⟩ with
x^ to get
|α⟩ is
equivalent to multiplying
ψ(x)=⟨x|ψ⟩ by
x to
get
α(x)=⟨x|α⟩. We have a definition of
x^ in the
coordinate representation.
[We have not bothered to try to get the x-representation of the
matrix of the operator x^. If you are ambitious you can try to
show that
⟨x|x^|x′⟩=xδ(x−x′).(20.39)
You can then work out the amusing result that
x^|x⟩=x|x⟩.(20.40)
The operator
x^ has the interesting property that when it works on
the base states
|x⟩ it is equivalent to multiplying by
x.]
Do you want to know the average value of x2? It is
⟨x2⟩av=∫ψ∗(x)x2ψ(x)dx.(20.41)
Or, if you prefer you can write
⟨x2⟩av=⟨ψ|α′⟩
with
|α′⟩=x^2|ψ⟩.(20.42)
By
x^2 we mean
x^x^—the two operators are used one after
the other. With the second form you can calculate
⟨x2⟩av, using
any representation (base-states) you wish. If you want the average
of
xn, or of any polynomial in
x, you can see how to get it.
20–5The momentum operator
Now we would like to calculate the mean momentum of an
electron—again, we’ll stick to one dimension. Let P(p)dp be the
probability that a measurement will give a momentum between p
and p+dp. Then
⟨p⟩av=∫pP(p)dp.(20.43)
Now we let
⟨p|ψ⟩ be the amplitude that the
state
|ψ⟩ is in a definite momentum state
|p⟩. This is the
same amplitude we called
⟨momp|ψ⟩ in
Section
16–3 and is a function of
p just
as
⟨x|ψ⟩ is a function of
x. There we chose to normalize the
amplitude so that
P(p)=12πℏ|⟨p|ψ⟩|2.(20.44)
We have, then,
⟨p⟩av=∫⟨ψ|p⟩p⟨p|ψ⟩dp2πℏ.(20.45)
The form is quite similar to what we had for
⟨x⟩av.
If we want, we can play exactly the same game we did
with ⟨x⟩av. First, we can write the integral above as
∫⟨ψ|p⟩⟨p|β⟩dp2πℏ.(20.46)
You should now recognize this equation as just the expanded form of
the amplitude
⟨ψ|β⟩—expanded in terms of the base
states of definite momentum. From Eq. (
20.45) the
state
|β⟩ is defined
in the momentum representation by
⟨p|β⟩=p⟨p|ψ⟩(20.47)
That is, we can now write
⟨p⟩av=⟨ψ|β⟩(20.48)
with
|β⟩=p^|ψ⟩,(20.49)
where the operator
p^ is defined in terms of the
p-representation by Eq. (
20.47).
[Again, you can if you wish show that the matrix form of p^ is
⟨p|p^|p′⟩=pδ(p−p′),(20.50)
and that
p^|p⟩=p|p⟩.(20.51)
It works out the same as for
x.]
Now comes an interesting question. We can write ⟨p⟩av, as we have done in
Eqs. (20.45) and (20.48), and we know the
meaning of the operator p^ in the momentum representation. But how
should we interpret p^ in the coordinate representation? That is what
we will need to know if we have some wave function ψ(x), and we want to
compute its average momentum. Let’s make clear what we mean. If we start by
saying that ⟨p⟩av is given by Eq. (20.48), we can expand that
equation in terms of the p-representation to get back to
Eq. (20.46). If we are given the p-description of the
state—namely the amplitude ⟨p|ψ⟩, which is an algebraic function
of the momentum p—we can get ⟨p|β⟩ from
Eq. (20.47) and proceed to evaluate the integral. The question now
is: What do we do if we are given a description of the state in the
x-representation, namely the wave function ψ(x)=⟨x|ψ⟩?
Well, let’s start by expanding Eq. (20.48) in the
x-representation. It is
⟨p⟩av=∫⟨ψ|x⟩⟨x|β⟩dx.(20.52)
Now, however, we need to know what the state
|β⟩ is in the
x-representation. If we can find it, we can carry out the
integral. So our problem is to find the
function
β(x)=⟨x|β⟩.
We can find it in the following way. In Section 16–3 we
saw how ⟨p|β⟩ was related
to ⟨x|β⟩. According to Eq. (16.24),
⟨p|β⟩=∫e−ipx/ℏ⟨x|β⟩dx.(20.53)
If we know
⟨p|β⟩ we can solve this equation
for
⟨x|β⟩. What we want, of course, is to express the result
somehow in terms of
ψ(x)=⟨x|ψ⟩, which we are assuming
to be known. Suppose we start with Eq. (
20.47) and
again use Eq. (
16.24) to write
⟨p|β⟩=p⟨p|ψ⟩=p∫e−ipx/ℏψ(x)dx.(20.54)
Since the integral is over
x we can put the
p inside the integral
and write
⟨p|β⟩=∫e−ipx/ℏpψ(x)dx.(20.55)
Compare this with (
20.53). You would say that
⟨x|β⟩ is equal to
pψ(x). No, No! The wave
function
⟨x|β⟩=β(x) can depend only on
x—not
on
p. That’s the whole problem.
However, some ingenious fellow discovered that the integral
in (20.55) could be integrated by parts. The derivative
of e−ipx/ℏ with respect to x is (−i/ℏ)pe−ipx/ℏ, so
the integral in (20.55) is equivalent to
−ℏi∫ddx(e−ipx/ℏ)ψ(x)dx.
If we integrate by parts, it becomes
−ℏi[e−ipx/ℏψ(x)]+∞−∞+ℏi∫e−ipx/ℏdψdxdx.
So long as we are considering bound states, so that
ψ(x) goes to
zero at
x=±∞, the bracket is zero and we have
⟨p|β⟩=ℏi∫e−ipx/ℏdψdxdx.(20.56)
Now compare this result with Eq. (
20.53). You
see that
⟨x|β⟩=ℏiddxψ(x).(20.57)
We have the necessary piece to be able to complete
Eq. (
20.52). The answer is
⟨p⟩av=∫ψ∗(x)ℏiddxψ(x)dx.(20.58)
We have found how Eq. (
20.48) looks in the coordinate
representation.
Now you should begin to see an interesting pattern developing. When we
asked for the average energy of the state |ψ⟩ we said it was
⟨E⟩av=⟨ψ|ϕ⟩, with |ϕ⟩=H^|ψ⟩.
The same thing is written in the coordinate world as
⟨E⟩av=∫ψ∗(x)ϕ(x)dx, with ϕ(x)=H^ψ(x).
Here
H^ is an
algebraic operator which works on a
function of
x. When we asked about the average value of
x, we
found that it could also be written
⟨x⟩av=⟨ψ|α⟩, with |α⟩=x^|ψ⟩.
In the coordinate world the corresponding equations are
⟨x⟩av=∫ψ∗(x)α(x)dx, with α(x)=xψ(x).
When we asked about the average value of
p, we wrote
⟨p⟩av=⟨ψ|β⟩, with |β⟩=p^|ψ⟩.
In the coordinate world the equivalent equations were
⟨p⟩av=∫ψ∗(x)β(x)dx, with β(x)=ℏiddxψ(x).
In each of our three examples we start with the state
|ψ⟩ and
produce another (hypothetical) state by a
quantum-mechanical
operator. In the coordinate representation we generate the
corresponding wave function by operating on the wave
function
ψ(x) with an
algebraic operator. There are the following
one-to-one correspondences (for one-dimensional problems):
H^x^p^x→H^=−ℏ22md2dx2+V(x),→x,→P^x=ℏi∂∂x.(20.59)
In this list, we have introduced the symbol
P^x for the
algebraic operator
(ℏ/i)∂/∂x:
P^x=ℏi∂∂x,(20.60)
and we have inserted the
x subscript on
P^ to remind you that
we have been working only with the
x-component of momentum.
You can easily extend the results to three dimensions. For the other
components of the momentum,
p^yp^z→P^y=ℏi∂∂y,→P^z=ℏi∂∂z.
If you want, you can even think of an operator of the
vector
momentum and write
p^→P^=ℏi(ex∂∂x+ey∂∂y+ez∂∂z),
where
ex,
ey, and
ez are the unit vectors in the
three directions. It looks even more elegant if we write
p^→P^=ℏi∇.(20.61)
Our general result is that for at least some quantum-mechanical
operators, there are corresponding algebraic operators in the
coordinate representation. We summarize our results so far—extended
to three dimensions—in Table 20–1. For each operator
we have the two equivalent forms:
|ϕ⟩=A^|ψ⟩(20.62)
or
ϕ(r)=A^ψ(r).(20.63)
Table 20–1
Physical Quantity |
Operator |
Coordinate Form |
Energytum |
H^ |
H^=−ℏ22m∇2+V(r) |
Positionum |
x^ |
$H^= x |
|
y^ |
$H^= y |
|
z^ |
$H^= z |
Momentum |
p^x |
P^x=ℏi∂∂x |
|
p^y |
P^y=ℏi∂∂y |
|
p^z |
P^z=ℏi∂∂z |
We will now give a few illustrations of the use of these ideas. The
first one is just to point out the relation between P^
and H^. If we use P^x twice, we get
P^xP^x=−ℏ2∂2∂x2.
This means that we can write the equality
H^=12m{P^xP^x+P^yP^y+P^zP^z}+V(r).
Or, using the vector notation,
H^=12mP^⋅P^+V(r).(20.64)
(In an algebraic operator, any term without the operator
symbol (
^) means just a straight multiplication.) This equation
is nice because it’s easy to remember if you haven’t forgotten your
classical physics. Everyone knows that the energy is
(nonrelativistically) just the kinetic energy
p2/2m plus the
potential energy, and
H^ is the operator of the total energy.
This result has impressed people so much that they try to teach
students all about classical physics before quantum mechanics. (We
think differently!) But such parallels are often misleading. For one
thing, when you have operators, the order of various factors is
important; but that is not true for the factors in a classical
equation.
In Chapter 17 we defined an operator p^x in terms
of the displacement operator D^x by [see Eq. (17.27)]
|ψ′⟩=D^x(δ)|ψ⟩=(1+iℏp^xδ)|ψ⟩,(20.65)
where
δ is a
small displacement. We should show you that
this is equivalent to our new definition. According to what we have
just worked out, this equation should mean the same as
ψ′(x)=ψ(x)+∂ψ∂xδ.
But the right-hand side is just the Taylor expansion
of
ψ(x+δ), which is certainly what you get if you displace the
state to the left by
δ (or shift the coordinates to the right
by the same amount). Our two definitions of
p^ agree!
Let’s use this fact to show something else. Suppose we have a bunch of
particles which we label 1, 2, 3, …, in some complicated
system. (To keep things simple we’ll stick to one dimension.) The wave
function describing the state is a function of all the coordinates
x1, x2, x3, … We can write it
as ψ(x1,x2,x3,…). Now displace the system (to the left)
by δ. The new wave function
ψ′(x1,x2,x3,…)=ψ(x1+δ,x2+δ,x3+δ,…)
can be written as
ψ′(x1,x2,x3,…)=ψ(x1,x2,x3,…)+{δ∂ψ∂x1+δ∂ψ∂x2+δ∂ψ∂x3+⋯}.(20.66)
According to Eq. (
20.65) the operator of the momentum
of the state
|ψ⟩ (let’s call it the
total momentum) is
equal to
P^total=ℏi{∂∂x1+∂∂x2+∂∂x3+⋯}.
But this is just the same as
P^total=P^x1+P^x2+P^x3+⋯.(20.67)
The operators of momentum obey the rule that the total momentum is the
sum of the momenta of all the parts. Everything holds together nicely,
and many of the things we have been saying are consistent with each
other.
20–6Angular momentum
Let’s for fun look at another operation—the operation of orbital
angular momentum. In Chapter 17 we defined an
operator J^z in terms of R^z(ϕ), the operator of a rotation by the
angle ϕ about the z-axis. We consider here a system described
simply by a single wave function ψ(r), which is a function of
coordinates only, and does not take into account the fact that the
electron may have its spin either up or down. That is, we want for the
moment to disregard intrinsic angular momentum and think about
only the orbital part. To keep the distinction clear, we’ll
call the orbital operator L^z, and define it in terms of the
operator of a rotation by an infinitesimal angle ϵ by
R^z(ϵ)|ψ⟩=(1+iℏϵL^z)|ψ⟩.
(Remember, this definition applies only to a state
|ψ⟩ which
has no internal spin variables, but depends only on the
coordinates
r=x,y,z.) If we look at the state
|ψ⟩ in a new
coordinate system, rotated about the
z-axis by the small
angle
ϵ, we see a new state
|ψ′⟩=R^z(ϵ)|ψ⟩.
If we choose to describe the state |ψ⟩ in the coordinate
representation—that is, by its wave function ψ(r), we would
expect to be able to write
ψ′(r)=(1+iℏϵL^z)ψ(r).(20.68)
What is
L^z? Well, a point
P at
x and
y in the
new
coordinate system (really
x′ and
y′, but we will drop the primes)
was formerly at
x−ϵy and
y+ϵx, as you can see from
Fig.
20–2. Since the amplitude for the electron to be at
P isn’t
changed by the rotation of the coordinates we can write
ψ′(x,y,z)=ψ(x−ϵy,y+ϵx,z)=ψ(x,y,z)−ϵy∂ψ∂x+ϵx∂ψ∂y
(remembering that
ϵ is a small angle). This means that
L^z=ℏi(x∂∂y−y∂∂x).(20.69)
That’s our answer. But notice. It is equivalent to
L^z=xP^y−yP^x.(20.70)
Returning to our quantum-mechanical operators, we can write
L^z=xp^y−yp^x.(20.71)
This formula is easy to remember because it looks like the familiar
formula of classical mechanics; it is the
z-component of
L=r×p.(20.72)
One of the fun parts of this operator business is that many classical
equations get carried over into a quantum-mechanical form. Which ones
don’t? There had better be some that don’t come out right, because if
everything did, then there would be nothing different about quantum
mechanics. There would be no new physics. Here is one equation which
is different. In classical physics
xpx−pxx=0.
What is it in quantum mechanics?
x^p^x−p^xx^=?
Let’s work it out in the
x-representation. So that we’ll know what
we are doing we put in some wave function
ψ(x). We have
xP^xψ(x)−P^xxψ(x),
or
xℏi∂∂xψ(x)−ℏi∂∂xxψ(x).
Remember now that the derivatives operate on everything to the
right. We get
xℏi∂ψ∂x−ℏiψ(x)−ℏix∂ψ∂x=−ℏiψ(x).(20.73)
The answer is
not zero. The whole operation is equivalent
simply to multiplication by
−ℏ/i:
x^p^x−p^xx^=−ℏi.(20.74)
If Planck’s constant were zero,
the classical and quantum results would be the same, and there would be
no quantum mechanics to learn!
Incidentally, if any two operators A^ and B^, when taken
together like this:
A^B^−B^A^,
do
not give zero, we say that “the operators do not commute.”
And an equation such as (
20.74) is called a “commutation
rule.” You can see that the commutation rule
for
px and
y is
p^xy^−y^p^x=0.
There is another very important commutation rule that has to do with
angular momenta. It is
L^xL^y−L^yL^x=iℏL^z.(20.75)
You can get some practice with
x^ and
p^ operators by proving
it for yourself.
It is interesting to notice that operators which do not commute can
also occur in classical physics. We have already seen this when we
have talked about rotation in space. If you rotate something, such as
a book, by 90∘ around x and then 90∘ around y, you
get something different from rotating first by 90∘ around y
and then by 90∘ around x. It is, in fact, just this property
of space that is responsible for Eq. (20.75).
20–7The change of averages with time
Now we want to show you something else. How do averages change with
time? Suppose for the moment that we have an operator A^, which
does not itself have time in it in any obvious way. We mean an
operator like x^ or p^. (We exclude things like, say, the
operator of some external potential that was being varied with time,
such as V(x,t).) Now suppose we calculate ⟨A⟩av, in some
state |ψ⟩, which is
⟨A⟩av=⟨ψ|A^|ψ⟩.(20.76)
How will
⟨A⟩av, depend on time? Why should it? One reason might be
that the operator itself depended explicitly on time—for instance,
if it had to do with a time-varying potential like
V(x,t). But even
if the operator does not depend on
t, say, for example, the
operator
A^=x^, the corresponding average may depend on time. Certainly
the average position of a particle could be moving. How does such a
motion come out of Eq. (
20.76) if
A^ has no time
dependence? Well, the state
|ψ⟩ might be changing with
time. For nonstationary states we have often shown a time dependence
explicitly by writing a state as
|ψ(t)⟩. We want to show that
the rate of change of
⟨A⟩av, is given by a new operator we will
call
A˙^. Remember that
A^ is an operator, so that putting a
dot over the
A does not here mean taking the time derivative, but is
just a way of writing a
new operator
A˙^ which is defined
by
ddt⟨A⟩av=⟨ψ|A˙^|ψ⟩.(20.77)
Our problem is to find the operator
A˙^.
First, we know that the rate of change of a state is given by the
Hamiltonian. Specifically,
iℏddt|ψ(t)⟩=H^|ψ(t)⟩.(20.78)
This is just the abstract way of writing our original definition of
the Hamiltonian:
iℏdCidt=∑jHijCj.(20.79)
If we take the complex conjugate of Eq. (
20.78), it is
equivalent to
−iℏddt⟨ψ(t)|=⟨ψ(t)|H^.(20.80)
Next, see what happens if we take the derivatives with respect to
t
of Eq. (
20.76). Since each
ψ depends on
t, we
have
ddt⟨A⟩av=(ddt⟨ψ|)A^|ψ⟩+⟨ψ|A^(ddt|ψ⟩).(20.81)
Finally, using the two equations in (
20.78)
and (
20.80) to replace the derivatives, we get
ddt⟨A⟩av=iℏ{⟨ψ|H^A^|ψ⟩−⟨ψ|A^H^|ψ⟩}.
This equation is the same as
ddt⟨A⟩av=iℏ⟨ψ|H^A^−A^H^|ψ⟩.
Comparing this equation with Eq. (
20.77), you see that
A˙^=iℏ(H^A^−A^H^).(20.82)
That is our interesting proposition, and it is true for any operator
A^.
Incidentally, if the operator A^ should itself be time
dependent, we would have had
A˙^=iℏ(H^A^−A^H^)+∂A^∂t.(20.83)
Let us try out Eq. (20.82) on some example to see
whether it really makes sense. For instance, what operator corresponds
to x˙^? We say it should be
x˙^=iℏ(H^x^−x^H^).(20.84)
What is this? One way to find out is to work it through in the
coordinate representation using the algebraic operator for
H^.
In this representation the commutator is
H^x−xH^={−ℏ22md2dx2+V(x)}x−x{−ℏ22md2dx2+V(x)}.
If you operate with this on any wave function
ψ(x) and work out
all of the derivatives where you can, you end up after a little work
with
−ℏ2mdψdx.
But this is just the same as
−iℏmP^xψ,
so we find that
H^x^−x^H^=−iℏmp^x(20.85)
or that
x˙^=p^xm.(20.86)
A pretty result. It means that if the mean value of
x is changing
with time the drift of the center of gravity is the same as the mean
momentum divided by
m. Exactly like classical mechanics.
Another example. What is the rate of change of the average momentum of
a state? Same game. Its operator is
p˙^=iℏ(H^p^−p^H^).(20.87)
Again you can work it out in the
x representation. Remember that
p^ becomes
d/dx, and this means that you will be taking the
derivative of the potential energy
V (in the
H^)—but only
in the second term. It turns out that it is the only term which does
not cancel, and you find that
H^P^−P^H^=iℏdVdx
or that
p˙^=−dVdx.(20.88)
Again the classical result. The right-hand side is the force, so we
have derived Newton’s law! But remember—these are the laws for the
operators which give the
average quantities. They do not
describe what goes on in detail inside an atom.
Quantum mechanics has the essential difference that p^x^ is not
equal to x^p^. They differ by a little bit—by the small
number iℏ. But the whole wondrous complications of interference, waves,
and all, result from the little fact that x^p^−p^x^ is not
quite zero.
The history of this idea is also interesting. Within a period of a few
months in 1926, Heisenberg and
Schrödinger
independently found
correct laws to describe atomic mechanics.
Schrödinger
invented his wave function ψ(x) and found his
equation.
Heisenberg, on the
other hand, found that nature could be described by classical
equations, except that xp−px should be equal to iℏ, which he
could make happen by defining them in terms of special kinds of
matrices. In our language he was using the energy-representation, with
its matrices. Both Heisenberg’s
matrix
algebra and
Schrödinger’s differential equation
explained the hydrogen atom. A few months later
Schrödinger was able to show
that the two theories were equivalent—as we have seen here. But the
two different mathematical forms of quantum mechanics were discovered
independently.