◄ ▲ ► A A A SUMMARY RECORDING
MATHJAX

http://www.feynmanlectures.caltech.edu/I_01.html

If it does not open, or only shows you this message again, then please let us know:

• which browser you are using (including version #)
• which operating system you are using (including version #)

This type of problem is rare, and there's a good chance it can be fixed if we have some clues about the cause. So, if you can, after enabling javascript, clearing the cache and disabling extensions, please open your browser's javascript console, load the page above, and if this generates any messages (particularly errors or warnings) on the console, then please make a copy (text or screenshot) of those messages and send them with the above-listed information to the email address given below.

By sending us information you will be helping not only yourself, but others who may be having similar problems accessing the online edition of The Feynman Lectures on Physics. Your time and consideration are greatly appreciated.

Best regards,
Mike Gottlieb
mg@feynmanlectures.info
Editor, The Feynman Lectures on Physics New Millennium Edition

## 20Operators

### 20–1Operations and operators

All the things we have done so far in quantum mechanics could be handled with ordinary algebra, although we did from time to time show you some special ways of writing quantum-mechanical quantities and equations. We would like now to talk some more about some interesting and useful mathematical ways of describing quantum-mechanical things. There are many ways of approaching the subject of quantum mechanics, and most books use a different approach from the one we have taken. As you go on to read other books you might not see right away the connections of what you will find in them to what we have been doing. Although we will also be able to get a few useful results, the main purpose of this chapter is to tell you about some of the different ways of writing the same physics. Knowing them you should be able to understand better what other people are saying. When people were first working out classical mechanics they always wrote all the equations in terms of $x$-, $y$-, and $z$-components. Then someone came along and pointed out that all of the writing could be made much simpler by inventing the vector notation. It’s true that when you come down to figuring something out you often have to convert the vectors back to their components. But it’s generally much easier to see what’s going on when you work with vectors and also easier to do many of the calculations. In quantum mechanics we were able to write many things in a simpler way by using the idea of the “state vector.” The state vector $\ket{\psi}$ has, of course, nothing to do with geometric vectors in three dimensions but is an abstract symbol that stands for a physical state, identified by the “label,” or “name,” $\psi$. The idea is useful because the laws of quantum mechanics can be written as algebraic equations in terms of these symbols. For instance, our fundamental law that any state can be made up from a linear combination of base states is written as $$\label{Eq:III:20:1} \ket{\psi}=\sum_iC_i\,\ket{i},$$ where the $C_i$ are a set of ordinary (complex) numbers—the amplitudes $C_i=\braket{i}{\psi}$—while $\ket{1}$, $\ket{2}$, $\ket{3}$, and so on, stand for the base states in some base, or representation.

If you take some physical state and do something to it—like rotating it, or like waiting for the time $\Delta t$—you get a different state. We say, “performing an operation on a state produces a new state.” We can express the same idea by an equation: $$\label{Eq:III:20:2} \ket{\phi}=\Aop\,\ket{\psi}.$$ An operation on a state produces another state. The operator $\Aop$ stands for some particular operation. When this operation is performed on any state, say $\ket{\psi}$, it produces some other state $\ket{\phi}$.

What does Eq. (20.2) mean? We define it this way. If you multiply the equation by $\bra{i}$ and expand $\ket{\psi}$ according to Eq. (20.1), you get $$\label{Eq:III:20:3} \braket{i}{\phi}=\sum_j\bracket{i}{\Aop}{j}\braket{j}{\psi}.$$ (The states $\ket{j}$ are from the same set as $\ket{i}$.) This is now just an algebraic equation. The numbers $\braket{i}{\phi}$ give the amount of each base state you will find in $\ket{\phi}$, and it is given in terms of a linear superposition of the amplitudes $\braket{j}{\psi}$ that you find $\ket{\psi}$ in each base state. The numbers $\bracket{i}{\Aop}{j}$ are just the coefficients which tell how much of $\braket{j}{\psi}$ goes into each sum. The operator $\Aop$ is described numerically by the set of numbers, or “matrix,” $$\label{Eq:III:20:4} A_{ij}\equiv\bracket{i}{\Aop}{j}.$$

So Eq. (20.2) is a high-class way of writing Eq. (20.3). Actually it is a little more than that; something more is implied. In Eq. (20.2) we do not make any reference to a set of base states. Equation (20.3) is an image of Eq. (20.2) in terms of some set of base states. But, as you know, you may use any set you wish. And this idea is implied in Eq. (20.2). The operator way of writing avoids making any particular choice. Of course, when you want to get definite you have to choose some set. When you make your choice, you use Eq. (20.3). So the operator equation (20.2) is a more abstract way of writing the algebraic equation (20.3). It’s similar to the difference between writing \begin{equation*} \FLPc=\FLPa\times\FLPb \end{equation*} instead of \begin{aligned} c_x&=a_yb_z-a_zb_y,\\[.5ex] c_y&=a_zb_x-a_xb_z,\\[.5ex] c_z&=a_xb_y-a_yb_x. \end{aligned} The first way is much handier. When you want results, however, you will eventually have to give the components with respect to some set of axes. Similarly, if you want to be able to say what you really mean by $\Aop$, you will have to be ready to give the matrix $A_{ij}$ in terms of some set of base states. So long as you have in mind some set $\ket{i}$, Eq. (20.2) means just the same as Eq. (20.3). (You should remember also that once you know a matrix for one particular set of base states you can always calculate the corresponding matrix that goes with any other base. You can transform the matrix from one “representation” to another.)

The operator equation in (20.2) also allows a new way of thinking. If we imagine some operator $\Aop$, we can use it with any state $\ket{\psi}$ to create a new state $\Aop\,\ket{\psi}$. Sometimes a “state” we get this way may be very peculiar—it may not represent any physical situation we are likely to encounter in nature. (For instance, we may get a state that is not normalized to represent one electron.) In other words, we may at times get “states” that are mathematically artificial. Such artificial “states” may still be useful, perhaps as the mid-point of some calculation.

We have already shown you many examples of quantum-mechanical operators. We have had the rotation operator $\Rop_y(\theta)$ which takes a state $\ket{\psi}$ and produces a new state, which is the old state as seen in a rotated coordinate system. We have had the parity (or inversion) operator $\Pop$, which makes a new state by reversing all coordinates. We have had the operators $\sigmaop_x$, $\sigmaop_y$, and $\sigmaop_z$ for spin one-half particles.

The operator $\Jop_z$ was defined in Chapter 17 in terms of the rotation operator for a small angle $\epsilon$. $$\label{Eq:III:20:5} \Rop_z(\epsilon)=1+\frac{i}{\hbar}\,\epsilon\,\Jop_z.$$ This just means, of course, that $$\label{Eq:III:20:6} \Rop_z(\epsilon)\,\ket{\psi}=\ket{\psi}+ \frac{i}{\hbar}\,\epsilon\,\Jop_z\,\ket{\psi}.$$ In this example, $\Jop_z\,\ket{\psi}$ is $\hbar/i\epsilon$ times the state you get if you rotate $\ket{\psi}$ by the small angle $\epsilon$ and then subtract the original state. It represents a “state” which is the difference of two states.

One more example. We had an operator $\pop_x$—called the momentum operator ($x$-component) defined in an equation like (20.6). If $\Dop_x(L)$ is the operator which displaces a state along $x$ by the distance $L$, then $\pop_x$ is defined by $$\label{Eq:III:20:7} \Dop_x(\delta)=1+\frac{i}{\hbar}\,\delta\pop_x,$$ where $\delta$ is a small displacement. Displacing the state $\ket{\psi}$ along $x$ by a small distance $\delta$ gives a new state $\ket{\psi'}$. We are saying that this new state is the old state plus a small new piece \begin{equation*} \frac{i}{\hbar}\,\delta\pop_x\,\ket{\psi}. \end{equation*}

The operators we are talking about work on a state vector like $\ket{\psi}$, which is an abstract description of a physical situation. They are quite different from algebraic operators which work on mathematical functions. For instance, $d/dx$ is an “operator” that works on $f(x)$ by changing it to a new function $f'(x)=df/dx$. Another example is the algebraic operator $\nabla^2$. You can see why the same word is used in both cases, but you should keep in mind that the two kinds of operators are different. A quantum-mechanical operator $\Aop$ does not work on an algebraic function, but on a state vector like $\ket{\psi}$. Both kinds of operators are used in quantum mechanics and often in similar kinds of equations, as you will see a little later. When you are first learning the subject it is well to keep the distinction always in mind. Later on, when you are more familiar with the subject, you will find that it is less important to keep any sharp distinction between the two kinds of operators. You will, indeed, find that most books generally use the same notation for both!

We’ll go on now and look at some useful things you can do with operators. But first, one special remark. Suppose we have an operator $\Aop$ whose matrix in some base is $A_{ij}\equiv\bracket{i}{\Aop}{j}$. The amplitude that the state $\Aop\,\ket{\psi}$ is also in some other state $\ket{\phi}$ is $\bracket{\phi}{\Aop}{\psi}$. Is there some meaning to the complex conjugate of this amplitude? You should be able to show that $$\label{Eq:III:20:8} \bracket{\phi}{\Aop}{\psi}\cconj=\bracket{\psi}{\Aop\adj}{\phi},$$ where $\Aop\adj$ (read “A dagger”) is an operator whose matrix elements are $$\label{Eq:III:20:9} A_{ij}\adj=(A_{ji})\cconj.$$ To get the $i,j$ element of $A\adj$ you go to the $j,i$ element of $A$ (the indexes are reversed) and take its complex conjugate. The amplitude that the state $\Aop\adj\,\ket{\phi}$ is in $\ket{\psi}$ is the complex conjugate of the amplitude that $\Aop\,\ket{\psi}$ is in $\ket{\phi}$. The operator $\Aop\adj$ is called the “Hermitian adjoint” of $\Aop$. Many important operators of quantum mechanics have the special property that when you take the Hermitian adjoint, you get the same operator back. If $\Bop$ is such an operator, then \begin{equation*} \Bop\adj=\Bop, \end{equation*} and it is called a “self-adjoint” or “Hermitian,” operator.

### 20–2Average energies

So far we have reminded you mainly of what you already know. Now we would like to discuss a new question. How would you find the average energy of a system—say, an atom? If an atom is in a particular state of definite energy and you measure the energy, you will find a certain energy $E$. If you keep repeating the measurement on each one of a whole series of atoms which are all selected to be in the same state, all the measurements will give $E$, and the “average” of your measurements will, of course, be just $E$.

Now, however, what happens if you make the measurement on some state $\ket{\psi}$ which is not a stationary state? Since the system does not have a definite energy, one measurement would give one energy, the same measurement on another atom in the same state would give a different energy, and so on. What would you get for the average of a whole series of energy measurements?

We can answer the question by projecting the state $\ket{\psi}$ onto the set of states of definite energy. To remind you that this is a special base set, we’ll call the states $\ket{\eta_i}$. Each of the states $\ket{\eta_i}$ has a definite energy $E_i$. In this representation, $$\label{Eq:III:20:10} \ket{\psi}=\sum_iC_i\,\ket{\eta_i}.$$ When you make an energy measurement and get some number $E_i$, you have found that the system was in the state $\eta_i$. But you may get a different number for each measurement. Sometimes you will get $E_1$, sometimes $E_2$, sometimes $E_3$, and so on. The probability that you observe the energy $E_1$ is just the probability of finding the system in the state $\ket{\eta_1}$, which is, of course, just the absolute square of the amplitude $C_1=\braket{\eta_1}{\psi}$. The probability of finding each of the possible energies $E_i$ is $$\label{Eq:III:20:11} P_i=\abs{C_i}^2.$$

How are these probabilities related to the mean value of a whole sequence of energy measurements? Let’s imagine that we get a series of measurements like this: $E_1$, $E_7$, $E_{11}$, $E_9$, $E_1$, $E_{10}$, $E_7$, $E_2$, $E_3$, $E_9$, $E_6$, $E_4$, and so on. We continue for, say, a thousand measurements. When we are finished we add all the energies and divide by one thousand. That’s what we mean by the average. There’s also a short-cut to adding all the numbers. You can count up how many times you get $E_1$, say that is $N_1$, and then count up the number of times you get $E_2$, call that $N_2$, and so on. The sum of all the energies is certainly just \begin{equation*} N_1E_1+N_2E_2+N_3E_3+\dotsb{}=\sum_iN_iE_i. \end{equation*} The average energy is this sum divided by the total number of measurements which is just the sum of all the $N_i$'s, which we can call $N$; $$\label{Eq:III:20:12} E_{\text{av}}=\frac{\sum_iN_iE_i}{N}.$$

We are almost there. What we mean by the probability of something happening is just the number of times we expect it to happen divided by the total number of tries. The ratio $N_i/N$ should—for large $N$—be very near to $P_i$, the probability of finding the state $\ket{\eta_i}$, although it will not be exactly $P_i$ because of the statistical fluctuations. Let’s write the predicted (or “expected”) average energy as $\av{E}$; then we can say that $$\label{Eq:III:20:13} \av{E}=\sum_iP_iE_i.$$ The same arguments apply for any measurement. The average value of a measured quantity $A$ should be equal to \begin{equation*} \av{A}=\sum_iP_iA_i, \end{equation*} where $A_i$ are the various possible values of the observed quantity, and $P_i$ is the probability of getting that value.

Let’s go back to our quantum-mechanical state $\ket{\psi}$. Its average energy is $$\label{Eq:III:20:14} \av{E}=\sum_i\abs{C_i}^2E_i=\sum_iC_i\cconj C_iE_i.$$ Now watch this trickery! First, we write the sum as $$\label{Eq:III:20:15} \sum_i\braket{\psi}{\eta_i}E_i\braket{\eta_i}{\psi}.$$ Next we treat the left-hand $\bra{\psi}$ as a common “factor.” We can take this factor out of the sum, and write it as \begin{equation*} \bra{\psi}\,\biggl\{\sum_i\ket{\eta_i}E_i\braket{\eta_i}{\psi}\biggr\}. \end{equation*} This expression has the form \begin{equation*} \braket{\psi}{\phi}, \end{equation*} where $\ket{\phi}$ is some “cooked-up” state defined by $$\label{Eq:III:20:16} \ket{\phi}=\sum_i\ket{\eta_i}E_i\braket{\eta_i}{\psi}.$$ It is, in other words, the state you get if you take each base state $\ket{\eta_i}$ in the amount $E_i\braket{\eta_i}{\psi}$.

Now remember what we mean by the states $\ket{\eta_i}$. They are supposed to be the stationary states—by which we mean that for each one, \begin{equation*} \Hop\,\ket{\eta_i}=E_i\,\ket{\eta_i}. \end{equation*} Since $E_i$ is just a number, the right-hand side is the same as $\ket{\eta_i}E_i$, and the sum in Eq. (20.16) is the same as \begin{equation*} \sum_i\Hop\,\ket{\eta_i}\braket{\eta_i}{\psi}. \end{equation*} Now $i$ appears only in the famous combination that contracts to unity, so \begin{equation*} \sum_i\Hop\,\ket{\eta_i}\braket{\eta_i}{\psi}= \Hop\sum_i\ket{\eta_i}\braket{\eta_i}{\psi}=\Hop\,\ket{\psi}. \end{equation*} Magic! Equation (20.16) is the same as $$\label{Eq:III:20:17} \ket{\phi}=\Hop\,\ket{\psi}.$$ The average energy of the state $\ket{\psi}$ can be written very prettily as $$\label{Eq:III:20:18} \av{E}=\bracket{\psi}{\Hop}{\psi}.$$ To get the average energy you operate on $\ket{\psi}$ with $\Hop$, and then multiply by $\bra{\psi}$. A simple result.

Our new formula for the average energy is not only pretty. It is also useful, because now we don’t need to say anything about any particular set of base states. We don’t even have to know all of the possible energy levels. When we go to calculate, we’ll need to describe our state in terms of some set of base states, but if we know the Hamiltonian matrix $H_{ij}$ for that set we can get the average energy. Equation (20.18) says that for any set of base states $\ket{i}$, the average energy can be calculated from $$\label{Eq:III:20:19} \av{E}=\sum_{ij}\braket{\psi}{i}\bracket{i}{\Hop}{j}\braket{j}{\psi},$$ where the amplitudes $\bracket{i}{\Hop}{j}$ are just the elements of the matrix $H_{ij}$.

Let’s check this result for the special case that the states $\ket{i}$ are the definite energy states. For them, $\Hop\,\ket{j}=E_j\,\ket{j}$, so $\bracket{i}{\Hop}{j}=E_j\,\delta_{ij}$ and \begin{equation*} \av{E}=\sum_{ij}\braket{\psi}{i}E_j\delta_{ij}\braket{j}{\psi}= \sum_iE_i\braket{\psi}{i}\braket{i}{\psi}, \end{equation*} which is right.

Equation (20.19) can, incidentally, be extended to other physical measurements which you can express as an operator. For instance, $\Lop_z$ is the operator of the $z$-component of the angular momentum $\FLPL$. The average of the $z$-component for the state $\ket{\psi}$ is \begin{equation*} \av{L_z}=\bracket{\psi}{\Lop_z}{\psi}. \end{equation*} One way to prove it is to think of some situation in which the energy is proportional to the angular momentum. Then all the arguments go through in the same way.

In summary, if a physical observable $A$ is related to a suitable quantum-mechanical operator $\Aop$, the average value of $A$ for the state $\ket{\psi}$ is given by $$\label{Eq:III:20:20} \av{A}=\bracket{\psi}{\Aop}{\psi}.$$ By this we mean that $$\label{Eq:III:20:21} \av{A}=\braket{\psi}{\phi},$$ with $$\label{Eq:III:20:22} \ket{\phi}=\Aop\,\ket{\psi}.$$

### 20–3The average energy of an atom

Suppose we want the average energy of an atom in a state described by a wave function $\psi(\FLPr)$; How do we find it? Let’s first think of a one-dimensional situation with a state $\ket{\psi}$ defined by the amplitude $\braket{x}{\psi}=\psi(x)$. We are asking for the special case of Eq. (20.19) applied to the coordinate representation. Following our usual procedure, we replace the states $\ket{i}$ and $\ket{j}$ by $\ket{x}$ and $\ket{x'}$, and change the sums to integrals. We get $$\label{Eq:III:20:23} \av{E}=\!\iint\braket{\psi}{x}\bracket{x}{\Hop}{x'} \braket{x'}{\psi}\,dx\,dx'\!.$$ This integral can, if we wish, be written in the following way: $$\label{Eq:III:20:24} \int\braket{\psi}{x}\braket{x}{\phi}\,dx,$$ with $$\label{Eq:III:20:25} \braket{x}{\phi}=\int\bracket{x}{\Hop}{x'}\braket{x'}{\psi}\,dx'.$$ The integral over $x'$ in (20.25) is the same one we had in Chapter 16—see Eq. (16.50) and Eq. (16.52)—and is equal to \begin{equation*} -\frac{\hbar^2}{2m}\,\frac{d^2}{dx^2}\,\psi(x)+V(x)\psi(x). \end{equation*} We can therefore write $$\label{Eq:III:20:26} \braket{x}{\phi}=\biggl\{\! -\frac{\hbar^2}{2m}\,\frac{d^2}{dx^2}+V(x)\biggr\}\psi(x).$$

Remember that $\braket{\psi}{x}=$ $\braket{x}{\psi}\cconj=$ $\psi\cconj(x)$; using this equality, the average energy in Eq. (20.23) can be written as $$\label{Eq:III:20:27} \av{E}=\!\int\!\psi\cconj(x)\biggl\{\! -\frac{\hbar^2}{2m}\frac{d^2}{dx^2}\!+\!V(x)\!\biggr\}\psi(x)\,dx.$$ Given a wave function $\psi(x)$, you can get the average energy by doing this integral. You can begin to see how we can go back and forth from the state-vector ideas to the wave-function ideas.

The quantity in the braces of Eq. (20.27) is an algebraic operator.1 We will write it as $\Hcalop$ \begin{equation*} \Hcalop=-\frac{\hbar^2}{2m}\,\frac{d^2}{dx^2}+V(x). \end{equation*} With this notation Eq. (20.23) becomes $$\label{Eq:III:20:28} \av{E}=\int\psi\cconj(x)\Hcalop\psi(x)\,dx.$$

The algebraic operator $\Hcalop$ defined here is, of course, not identical to the quantum-mechanical operator $\Hop$. The new operator works on a function of position $\psi(x)=\braket{x}{\psi}$ to give a new function of $x$, $\phi(x)=\braket{x}{\phi}$; while $\Hop$ operates on a state vector $\ket{\psi}$ to give another state vector $\ket{\phi}$, without implying the coordinate representation or any particular representation at all. Nor is $\Hcalop$ strictly the same as $\Hop$ even in the coordinate representation. If we choose to work in the coordinate representation, we would interpret $\Hop$ in terms of a matrix $\bracket{x}{\Hop}{x'}$ which depends somehow on the two “indices” $x$ and $x'$; that is, we expect—according to Eq. (20.25)—that $\braket{x}{\phi}$ is related to all the amplitudes $\braket{x}{\psi}$ by an integration. On the other hand, we find that $\Hcalop$ is a differential operator. We have already worked out in Section 16–5 the connection between $\bracket{x}{\Hop}{x'}$ and the algebraic operator $\Hcalop$.

We should make one qualification on our results. We have been assuming that the amplitude $\psi(x)=\braket{x}{\psi}$ is normalized. By this we mean that the scale has been chosen so that \begin{equation*} \int\abs{\psi(x)}^2\,dx=1; \end{equation*} so the probability of finding the electron somewhere is unity. If you should choose to work with a $\psi(x)$ which is not normalized you should write $$\label{Eq:III:20:29} \av{E}=\frac{\int\psi\cconj(x)\Hcalop\psi(x)\,dx} {\int\psi\cconj(x)\psi(x)\,dx}.$$ It’s the same thing.

Notice the similarity in form between Eq. (20.28) and Eq. (20.18). These two ways of writing the same result appear often when you work with the $x$-representation. You can go from the first form to the second with any $\Aop$ which is a local operator, where a local operator is one which in the integral \begin{equation*} \int\bracket{x}{\Aop}{x'}\braket{x'}{\psi}\,dx' \end{equation*} can be written as $\Acalop\psi(x)$, where $\Acalop$ is a differential algebraic operator. There are, however, operators for which this is not true. For them you must work with the basic equations in (20.21) and (20.22).

You can easily extend the derivation to three dimensions. The result is that2 $$\label{Eq:III:20:30} \av{E}=\int\psi\cconj(\FLPr)\Hcalop\psi(\FLPr)\,dV,$$ with $$\label{Eq:III:20:31} \Hcalop=-\frac{\hbar^2}{2m}\,\nabla^2+V(\FLPr),$$ and with the understanding that $$\label{Eq:III:20:32} \int\abs{\psi}^2\,dV=1.$$ The same equations can be extended to systems with several electrons in a fairly obvious way, but we won’t bother to write down the results.

With Eq. (20.30) we can calculate the average energy of an atomic state even without knowing its energy levels. All we need is the wave function. It’s an important law. We’ll tell you about one interesting application. Suppose you want to know the ground-state energy of some system—say the helium atom, but it’s too hard to solve Schrödinger’s equation for the wave function, because there are too many variables. Suppose, however, that you take a guess at the wave function—pick any function you like—and calculate the average energy. That is, you use Eq. (20.29)—generalized to three dimensions-to find what the average energy would be if the atom were really in the state described by this wave function. This energy will certainly be higher than the ground-state energy which is the lowest possible energy the atom can have.3 Now pick another function and calculate its average energy. If it is lower than your first choice you are getting closer to the true ground-state energy. If you keep on trying all sorts of artificial states you will be able to get lower and lower energies, which come closer and closer to the ground-state energy. If you are clever, you will try some functions which have a few adjustable parameters. When you calculate the energy it will be expressed in terms of these parameters. By varying the parameters to give the lowest possible energy, you are trying out a whole class of functions at once. Eventually you will find that it is harder and harder to get lower energies and you will begin to be convinced that you are fairly close to the lowest possible energy. The helium atom has been solved in just this way—not by solving a differential equation, but by making up a special function with a lot of adjustable parameters which are eventually chosen to give the lowest possible value for the average energy.

### 20–4The position operator

What is the average value of the position of an electron in an atom? For any particular state $\ket{\psi}$ what is the average value of the coordinate $x$? We’ll work in one dimension and let you extend the ideas to three dimensions or to systems with more than one particle: We have a state described by $\psi(x)$, and we keep measuring $x$ over and over again. What is the average? It is \begin{equation*} \int xP(x)\,dx, \end{equation*} where $P(x)\,dx$ is the probability of finding the electron in a little element $dx$ at $x$. Suppose the probability density $P(x)$ varies with $x$ as shown in Fig. 20–1. The electron is most likely to be found near the peak of the curve. The average value of $x$ is also somewhere near the peak. It is, in fact, just the center of gravity of the area under the curve.

We have seen earlier that $P(x)$ is just $\abs{\psi(x)}^2=\psi\cconj(x)\psi(x)$, so we can write the average of $x$ as $$\label{Eq:III:20:33} \av{x}=\int\psi\cconj(x)x\psi(x)\,dx.$$

Our equation for $\av{x}$ has the same form as Eq. (20.28). For the average energy, the energy operator $\Hcalop$ appears between the two $\psi$’s, for the average position there is just $x$. (If you wish you can consider $x$ to be the algebraic operator “multiply by $x$.”) We can carry the parallelism still further, expressing the average position in a form which corresponds to Eq. (20.18). Suppose we just write $$\label{Eq:III:20:34} \av{x}=\braket{\psi}{\alpha}$$ with $$\label{Eq:III:20:35} \ket{\alpha}=\xop\,\ket{\psi},$$ and then see if we can find the operator $\xop$ which generates the state $\ket{\alpha}$, which will make Eq. (20.34) agree with Eq. (20.33). That is, we must find a $\ket{\alpha}$, so that $$\label{Eq:III:20:36} \braket{\psi}{\alpha}=\av{x}= \int\braket{\psi}{x}x\braket{x}{\psi}\,dx.$$ First, let’s expand $\braket{\psi}{\alpha}$ in the $x$-representation. It is $$\label{Eq:III:20:37} \braket{\psi}{\alpha}=\int\braket{\psi}{x}\braket{x}{\alpha}\,dx.$$ Now compare the integrals in the last two equations. You see that in the $x$-representation $$\label{Eq:III:20:38} \braket{x}{\alpha}=x\braket{x}{\psi}.$$ Operating on $\ket{\psi}$ with $\xop$ to get $\ket{\alpha}$ is equivalent to multiplying $\psi(x)=\braket{x}{\psi}$ by $x$ to get $\alpha(x)=\braket{x}{\alpha}$. We have a definition of $\xop$ in the coordinate representation.4

[We have not bothered to try to get the $x$-representation of the matrix of the operator $\xop$. If you are ambitious you can try to show that $$\label{Eq:III:20:39} \bracket{x}{\xop}{x'}=x\,\delta(x-x').$$ You can then work out the amusing result that $$\label{Eq:III:20:40} \xop\,\ket{x}=x\,\ket{x}.$$ The operator $\xop$ has the interesting property that when it works on the base states $\ket{x}$ it is equivalent to multiplying by $x$.]

Do you want to know the average value of $x^2$? It is $$\label{Eq:III:20:41} \av{x^2}=\int\psi\cconj(x)x^2\psi(x)\,dx.$$ Or, if you prefer you can write $$\av{x^2}=\braket{\psi}{\alpha'}\notag$$ with $$\label{Eq:III:20:42} \ket{\alpha'}=\xop^2\,\ket{\psi}.$$ By $\xop^2$ we mean $\xop\xop$—the two operators are used one after the other. With the second form you can calculate $\av{x^2}$, using any representation (base-states) you wish. If you want the average of $x^n$, or of any polynomial in $x$, you can see how to get it.

### 20–5The momentum operator

Now we would like to calculate the mean momentum of an electron—again, we’ll stick to one dimension. Let $P(p)\,dp$ be the probability that a measurement will give a momentum between $p$ and $p+dp$. Then $$\label{Eq:III:20:43} \av{p}=\int p\,P(p)\,dp.$$ Now we let $\braket{p}{\psi}$ be the amplitude that the state $\ket{\psi}$ is in a definite momentum state $\ket{p}$. This is the same amplitude we called $\braket{\mom p}{\psi}$ in Section 16–3 and is a function of $p$ just as $\braket{x}{\psi}$ is a function of $x$. There we chose to normalize the amplitude so that $$\label{Eq:III:20:44} P(p)=\frac{1}{2\pi\hbar}\,\abs{\braket{p}{\psi}}^2.$$ We have, then, $$\label{Eq:III:20:45} \av{p}=\int\braket{\psi}{p}p\braket{p}{\psi}\,\frac{dp}{2\pi\hbar}.$$ The form is quite similar to what we had for $\av{x}$.

If we want, we can play exactly the same game we did with $\av{x}$. First, we can write the integral above as $$\label{Eq:III:20:46} \int\braket{\psi}{p}\braket{p}{\beta}\,\frac{dp}{2\pi\hbar}.$$ You should now recognize this equation as just the expanded form of the amplitude $\braket{\psi}{\beta}$—expanded in terms of the base states of definite momentum. From Eq. (20.45) the state $\ket{\beta}$ is defined in the momentum representation by $$\label{Eq:III:20:47} \braket{p}{\beta}=p\braket{p}{\psi}$$ That is, we can now write $$\label{Eq:III:20:48} \av{p}=\braket{\psi}{\beta}$$ with $$\label{Eq:III:20:49} \ket{\beta}=\pop\,\ket{\psi},$$ where the operator $\pop$ is defined in terms of the $p$-representation by Eq. (20.47).

[Again, you can if you wish show that the matrix form of $\pop$ is $$\label{Eq:III:20:50} \bracket{p}{\pop}{p'}=p\,\delta(p-p'),$$ and that $$\label{Eq:III:20:51} \pop\,\ket{p}=p\,\ket{p}.$$ It works out the same as for $x$.]

Now comes an interesting question. We can write $\av{p}$, as we have done in Eqs. (20.45) and (20.48), and we know the meaning of the operator $\pop$ in the momentum representation. But how should we interpret $\pop$ in the coordinate representation? That is what we will need to know if we have some wave function $\psi(x)$, and we want to compute its average momentum. Let’s make clear what we mean. If we start by saying that $\av{p}$ is given by Eq. (20.48), we can expand that equation in terms of the $p$-representation to get back to Eq. (20.46). If we are given the $p$-description of the state—namely the amplitude $\braket{p}{\psi}$, which is an algebraic function of the momentum $p$—we can get $\braket{p}{\beta}$ from Eq. (20.47) and proceed to evaluate the integral. The question now is: What do we do if we are given a description of the state in the $x$-representation, namely the wave function $\psi(x)=\braket{x}{\psi}$?

Well, let’s start by expanding Eq. (20.48) in the $x$-representation. It is $$\label{Eq:III:20:52} \av{p}=\int\braket{\psi}{x}\braket{x}{\beta}\,dx.$$ Now, however, we need to know what the state $\ket{\beta}$ is in the $x$-representation. If we can find it, we can carry out the integral. So our problem is to find the function $\beta(x)=\braket{x}{\beta}$.

We can find it in the following way. In Section 16–3 we saw how $\braket{p}{\beta}$ was related to $\braket{x}{\beta}$. According to Eq. (16.24), $$\label{Eq:III:20:53} \braket{p}{\beta}=\int e^{-ipx/\hbar}\braket{x}{\beta}\,dx.$$ If we know $\braket{p}{\beta}$ we can solve this equation for $\braket{x}{\beta}$. What we want, of course, is to express the result somehow in terms of $\psi(x)=\braket{x}{\psi}$, which we are assuming to be known. Suppose we start with Eq. (20.47) and again use Eq. (16.24) to write $$\label{Eq:III:20:54} \braket{p}{\beta}=p\braket{p}{\psi}= p\int e^{-ipx/\hbar}\psi(x)\,dx.$$ Since the integral is over $x$ we can put the $p$ inside the integral and write $$\label{Eq:III:20:55} \braket{p}{\beta}=\int e^{-ipx/\hbar}p\psi(x)\,dx.$$ Compare this with (20.53). You would say that $\braket{x}{\beta}$ is equal to $p\psi(x)$. No, No! The wave function $\braket{x}{\beta}=\beta(x)$ can depend only on $x$—not on $p$. That’s the whole problem.

However, some ingenious fellow discovered that the integral in (20.55) could be integrated by parts. The derivative of $e^{-ipx/\hbar}$ with respect to $x$ is $(-i/\hbar)pe^{-ipx/\hbar}$, so the integral in (20.55) is equivalent to \begin{equation*} -\frac{\hbar}{i}\int\ddt{}{x}\,(e^{-ipx/\hbar})\psi(x)\,dx. \end{equation*} If we integrate by parts, it becomes \begin{equation*} -\frac{\hbar}{i}\bigl[e^{-ipx/\hbar}\psi(x)\bigr]_{-\infty}^{+\infty}+ \frac{\hbar}{i}\int e^{-ipx/\hbar}\ddt{\psi}{x}\,dx. \end{equation*} So long as we are considering bound states, so that $\psi(x)$ goes to zero at $x=\pm\infty$, the bracket is zero and we have $$\label{Eq:III:20:56} \braket{p}{\beta}=\frac{\hbar}{i}\int e^{-ipx/\hbar}\ddt{\psi}{x}\,dx.$$ Now compare this result with Eq. (20.53). You see that $$\label{Eq:III:20:57} \braket{x}{\beta}=\frac{\hbar}{i}\,\ddt{}{x}\,\psi(x).$$ We have the necessary piece to be able to complete Eq. (20.52). The answer is $$\label{Eq:III:20:58} \av{p}=\!\int\!\psi\cconj(x)\,\frac{\hbar}{i}\,\ddt{}{x}\,\psi(x)\,dx.$$ We have found how Eq. (20.48) looks in the coordinate representation.

Now you should begin to see an interesting pattern developing. When we asked for the average energy of the state $\ket{\psi}$ we said it was \begin{equation*} \av{E}=\braket{\psi}{\phi}, \text{ with } \ket{\phi}=\Hop\,\ket{\psi}. \end{equation*} The same thing is written in the coordinate world as \begin{equation*} \av{E}=\!\int\!\psi\cconj(x)\phi(x)\,dx, \text{ with } \phi(x)=\Hcalop\psi(x). \end{equation*} Here $\Hcalop$ is an algebraic operator which works on a function of $x$. When we asked about the average value of $x$, we found that it could also be written \begin{equation*} \av{x}=\braket{\psi}{\alpha}, \text{ with } \ket{\alpha}=\xop\,\ket{\psi}. \end{equation*} In the coordinate world the corresponding equations are \begin{equation*} \av{x}=\!\int\!\psi\cconj(x)\alpha(x)\,dx, \text{ with } \alpha(x)=x\psi(x). \end{equation*} When we asked about the average value of $p$, we wrote \begin{equation*} \av{p}=\braket{\psi}{\beta}, \text{ with } \ket{\beta}=\pop\,\ket{\psi}. \end{equation*} In the coordinate world the equivalent equations were \begin{equation*} \av{p}=\!\int\!\psi\cconj(x)\beta(x)\,dx, \text{ with } \beta(x)=\frac{\hbar}{i}\,\ddt{}{x}\,\psi(x). \end{equation*} In each of our three examples we start with the state $\ket{\psi}$ and produce another (hypothetical) state by a quantum-mechanical operator. In the coordinate representation we generate the corresponding wave function by operating on the wave function $\psi(x)$ with an algebraic operator. There are the following one-to-one correspondences (for one-dimensional problems): \label{Eq:III:20:59} \begin{aligned} \Hop&\to\Hcalop=-\frac{\hbar^2}{2m}\,\frac{d^2}{dx^2}+V(x),\\[1pt] \xop&\to x,\\[1pt] \pop_x&\to\Pcalop_x=\frac{\hbar}{i}\,\ddp{}{x}. \end{aligned} In this list, we have introduced the symbol $\Pcalop_x$ for the algebraic operator $(\hbar/i)\ddpl{}{x}$: $$\label{Eq:III:20:60} \Pcalop_x=\frac{\hbar}{i}\,\ddp{}{x},$$ and we have inserted the $x$ subscript on $\Pcalop$ to remind you that we have been working only with the $x$-component of momentum.

You can easily extend the results to three dimensions. For the other components of the momentum, \begin{align*} \pop_y&\to\Pcalop_y=\frac{\hbar}{i}\,\ddp{}{y},\\[1ex] \pop_z&\to\Pcalop_z=\frac{\hbar}{i}\,\ddp{}{z}. \end{align*} If you want, you can even think of an operator of the vector momentum and write \begin{equation*} \pvecop\to\Pcalvecop=\frac{\hbar}{i}\biggl( \FLPe_x\,\ddp{}{x}+\FLPe_y\,\ddp{}{y}+\FLPe_z\,\ddp{}{z}\biggr), \end{equation*} where $\FLPe_x$, $\FLPe_y$, and $\FLPe_z$ are the unit vectors in the three directions. It looks even more elegant if we write $$\label{Eq:III:20:61} \pvecop\to\Pcalvecop=\frac{\hbar}{i}\,\FLPnabla.$$

Our general result is that for at least some quantum-mechanical operators, there are corresponding algebraic operators in the coordinate representation. We summarize our results so far—extended to three dimensions—in Table 20–1. For each operator we have the two equivalent forms:5 $$\label{Eq:III:20:62} \ket{\phi}=\Aop\,\ket{\psi}$$ or $$\label{Eq:III:20:63} \phi(\FLPr)=\Acalop\psi(\FLPr).$$

Table 20–1