Definition of Expectation Value (EV)

AI Thread Summary
Expectation value (EV) is defined for both discrete and continuous probability distributions, with specific equations for each case. For discrete distributions, the expectation value is calculated using the formula <Q> = ∑ Q_n p_n, where Q_n represents the eigenvalue and p_n the probability of measuring that state. In continuous distributions, the expectation value is derived from a normalized wave function, expressed as <Q> = ∫ ψ*(x) Q(x) ψ(x) dx. Variance is also discussed, highlighting its calculation in both discrete and continuous contexts, emphasizing the relationship between expectation values and variance. The discussion concludes with examples illustrating the calculation of expectation values for rolling dice and quantum mechanical properties like position and momentum.
Messages
19,787
Reaction score
10,739
Definition/Summary

A short introduction to expectation value is given, both for discrete and continuous cases.

Equations

For discrete probability distributions,

&lt;Q&gt; \ = \ \sum _n Q_n p_n

For continuous distributions specified by a normalized, real space wave-function \psi(x),

&lt; Q &gt; = \int _{\text{All space}} \psi^*(x)Q(x)\psi(x) dx

Extended explanation

NOTATION:

The notation < > comes from statistics, so it is a general notation which QM scientists borrowed.


DEFINITIONS:

The expectation value of an observable associated with an operator Q is defined as:

&lt;Q&gt; = \sum _n Q_n p_n

in the case of a discrete spectrum, where Q_n is the eigenvalue of Q for a state labeled by the index n, and p_n is the probability of measuring the system in this state.


DISCRETE DISTRIBUTIONS:

Variance in statistics, discrete case:
(\Delta A ) ^2 = \sum _n (A_n - &lt;A&gt;)^2 p_n ,
\sum _n p_n = 1 ,
&lt;A&gt; = \sum _n A_nP_n ,
&lt;&lt;A&gt;&gt; = &lt;A&gt;
&lt;A&gt; is just a number, we can thus show that:
(\Delta A ) ^2 = &lt;A^2&gt; + &lt;A&gt;^2
and
&lt;(\Delta A ) ^2&gt; = (\Delta A ) ^2. as an exercise, show this.

where \sum _n p_n = 1 and A_n is the outcome of the n'th value.


EXAMPLE:

As an exercise, let's find the expectation value <D>, of the outcome of rolling dice:

&lt;D&gt; = 1 \cdot \dfrac{1}{6} + 2 \cdot \dfrac{1}{6} + 3 \cdot \dfrac{1}{6} + 4 \cdot \dfrac{1}{6} + 5 \cdot \dfrac{1}{6} + 5 \cdot \dfrac{1}{6} = \dfrac{7}{2}
since each value has the equal probability of 1/6 .


CONTINUOUS DISTRIBUTIONS:

Now this was for the discrete case, in the continuous case:
&lt; Q &gt; = \int _{\text{All space}} f(x)Q(x)f(x) dx
where f^2(x) is the probability density distribution : \int f^2(x) dx = 1.

That was if the probability density distribution is real, for complex valued (such as quantum mechanical wave functions):
&lt; Q &gt; = \int _{\text{All space}} \psi^*(x)Q(x)\psi(x) dx
\int |\psi (x)|^2 dx = 1.

EXAMPLES:

Position:
&lt; x &gt; = \int _{\text{All space}} \psi^*(x)x\psi(x) dx = \int x|\psi (x)|^2 dx

Momentum:
&lt; p &gt; = \int _{\text{All space}} \psi^*(x)(-i\hbar\dfrac{d}{dx})\psi(x) dx

Now the variance is:
\Delta Q ^2 = &lt;(Q - &lt;Q&gt;)^2&gt; = &lt;Q^2&gt; - &lt;Q&gt;^2

* This entry is from our old Library feature. If you know who wrote it, please let us know so we can attribute a writer. Thanks!
 
Physics news on Phys.org
In QM, we write expectation value rather than expected value. The former sounds better as almost every textbook uses that term. Secondly, while writing bra-ket notation in the future insight article, it would be better if \langle and \rangle are used rather than < and >.

(You can delete this post once you make the necessary changes, as this would then become useless.)
 
So I know that electrons are fundamental, there's no 'material' that makes them up, it's like talking about a colour itself rather than a car or a flower. Now protons and neutrons and quarks and whatever other stuff is there fundamentally, I want someone to kind of teach me these, I have a lot of questions that books might not give the answer in the way I understand. Thanks
Back
Top