Understanding Projectors in Quantum Mechanics: A Mathematical Approach

  • Thread starter Thread starter Sammywu
  • Start date Start date
  • Tags Tags
    Projector Qm
Click For Summary
The discussion centers on the mathematical understanding of projectors in quantum mechanics, particularly their definition and properties. A projector is defined as an operator P that satisfies P^t = P and P^2 = P, with eigenvalues of 0 and/or 1. The conversation includes derivations related to projectors acting on states in Hilbert space, emphasizing the importance of normalization and orthogonality in these operations. Additionally, the participants explore whether projectors can serve as a basis for certain vector spaces and clarify that non-trivial projectors do not have inverses due to their filtering nature. Overall, the thread highlights the mathematical intricacies of projectors and their role in quantum mechanics.
  • #31
So, I just try to see whether QM is always of trace class.

Let [tex] M = \sum_n a_n P_{\psi_n} [/tex]
where
[tex] \sum_n a_n = 1 [/tex]

TR [tex] QM = \sum_n ( Q \sum_i a_i P_{\psi_i} \psi_n , \psi_n ) [/tex]
[tex] = \sum_n ( Q a_n \psi_n , \psi_n ) [/tex]
[tex] = \sum_n a_n ( Q \psi_n, \psi_n ) [/tex]

Assuming
[tex] b_n = | ( Q \psi_n, \psi_n ) | [/tex]
, whether QM is of trace class will depends on whether
[tex] \sum_n a_n b_n [/tex] converges.

So, if I can find a set of [tex] \psi_n [/tex] such that [tex] b_n = 2/a_n [/tex] , then I have a QM not of trace class.
 
Physics news on Phys.org
  • #32
Sammywu said:
Now to keep it easy, I pick a
[tex] \psi = ( | q - q_0 | e^{-(q-q_0)^2 })^{1/2} [/tex]
This is not easy.


[tex] ( \psi , \psi ) = 1 [/tex]
True.


Considering in the interval
[ q_0-1 , q_0 +1 ] ,
[tex] q^2 * | \psi |^2 >= ( q_0 - 1)^2 * 2 * 1/2 * 1/e [/tex]
How so? For q = qo the LHS is 0.

------
... The idea behind the proof will work. But for technical reasons the proof has failed. Why not really keep it easy and choose ψ like below?

ψ(q)

= 1 , q Є (qo, qo + 1)
= 0 , otherwise

Then (Qψ, Qψ) > qo2 .

------------------------------------

Define: ║φ║ = √(φ, φ)

Then: ║φ + ξ║ ≤ ║φ║ + ║ξ║

.... :surprise:

------------------------------------

So, I just try to see whether QM is always of trace class.
In the most general case, QM may or may not be.

In the following, you carry out the inquiry well:

So, I just try to see whether QM is always of trace class.

Let [tex] M = \sum_n a_n P_{\psi_n} [/tex]
where
[tex] \sum_n a_n = 1 [/tex]

TR [tex] QM = \sum_n ( Q \sum_i a_i P_{\psi_i} \psi_n , \psi_n ) [/tex]
[tex] = \sum_n ( Q a_n \psi_n , \psi_n ) [/tex]
[tex] = \sum_n a_n ( Q \psi_n, \psi_n ) [/tex]

Assuming
[tex] b_n = | ( Q \psi_n, \psi_n ) | [/tex]
, whether QM is of trace class will depends on whether
[tex] \sum_n a_n b_n [/tex] converges.
Remember also that M is a state. So we also have

0≤an≤1 as well as ∑nan=1 .

But still, this is not enough. In general the series can still diverge.

This tells us that our current definition for a state M is still too general. While all physical states do satisfy the definition of M, not all M's are physical states.
 
Last edited:
  • #33
Eye,

Note I just added an integral sign in front.

Considering in the interval
[ q_0-1 , q_0 +1 ] ,
[tex] \int_{q_0-1}^{q_0+1} q^2 * | \psi |^2 >= ( q_0 - 1)^2 * 2 * 1/2 * 1/e [/tex]

Any way, I agree your proof is much easier and quicker than mine.

I need to read your response more thoroughly.

It seems you used the extended triangular inequality; that was what I was thinking yesterday but can't get it proved and working.

I did branch out to think the issue of mixed state:

[tex] < Q , P_{\sum_n a_n \psi_n} > = TR \ QP_{\sum_n a_n \psi_n} = [/tex]
[tex] \sum_n ( QP_{\sum_i a_i \psi_i} \psi_n , \psi_n ) = \sum_n ( Q \sum_j a_j \overline{a_n} \psi_j , \psi_n ) = [/tex]
[tex] \sum_n \overline{a_n} ( \sum_j a_j Q \psi_j , \psi_n ) = \sum_n \overline{a_n} \sum_j a_j ( Q \psi_j , \psi_n ) [/tex]

[tex] <Q | M > = < Q , \sum_n a_n P_{\psi_n } > = TR \ Q \sum_n a_n P_{\psi_n} = [/tex]
[tex]\sum_n a_n ( Q P_{\psi_n} \psi_n , \psi_n ) = \sum_n a_n ( Q \psi_n , \psi_n ) [/tex]

This shows that even though [tex] P_{\sum_n a_n \psi_n} \ and \ \sum_n a_n P_{\psi_n } [/tex], but they do have the same expectation value.

Now, for [tex] P_\sum_n a_n \psi_n [/tex] to be a state, [tex] (\sum_n a_n \psi_n , \sum_n a_n \psi_n ) = \sum a_n \overline\a_n = \sum a_n^2 \ needs \ to \ be \ one [/tex] . This condition is different from the condition for M to be a state in that [tex] TR \ M = \sum_n (\sum_i a_i P_\psi_i \psi_n , \psi_n ) = \sum a_n needs \ to \ be \ one [/tex] .

I thought I might be able to show some Q wll have the same expectation values for the mixed state and the pure state. Apparently this is a little tedious than I thought.
 
Last edited:
  • #34
Eye,

The previous one is a little messy. I was trying to see whether the '//' will give me a new line.

So, I redo it here.

I did branch out to think the issue of mixed state:

[tex] < Q , P_{\sum_n a_n \psi_n} > = TR \ QP_{\sum_n a_n \psi_n} = [/tex]
[tex] \sum_n ( QP_{\sum_i a_i \psi_i} \psi_n , \psi_n ) = [/tex]
[tex] \sum_n ( Q \sum_j a_j \overline{a_n} \psi_j , \psi_n ) = [/tex]
[tex] \sum_n \overline{a_n} ( \sum_j a_j Q \psi_j , \psi_n ) = [/tex]
[tex] \sum_n \overline{a_n} \sum_j a_j ( Q \psi_j , \psi_n ) [/tex]

[tex] <Q | M > = \sum_n a_n ( Q \psi_n , \psi_n ) [/tex]

as we alreday showed in earlier one.


Now, for [tex] P_\sum_n a_n \psi_n [/tex] to be a state,
[tex] (\sum_n a_n \psi_n , \sum_n a_n \psi_n ) = [/tex]
[tex] \sum a_n \overline{a_n} = \sum a_n^2 [\tex]
needs to be one.
This condition is different from the condition for M to be a state in that
[tex] TR \ M = \sum_n (\sum_i a_i P_\psi_i \psi_n , \psi_n ) = [/tex]
[tex] \sum a_n [/tex]
needs to be one.

I thought I might be able to show some Qs wll have the same expectation values for the mixed state and the pure state. Apparently this is a little tedious than I thought.

Any way, why do you say I is unbounded in a infinite dimensional space?

[tex] ( I \psi , I \psi ) = ( \psi , \psi ) < 2 ( \psi , \psi ) [/tex]

whether the space is infinite or finite dimensional.

Thanks[/tex]
 
Last edited:
  • #35
Eye,

Changing my tactic, I gathered some facts:

Let
[tex] \psi = \sum b_i \psi_i [/tex]
where
[tex] b_i \overline{b_i} = a_i [/tex]
and
[tex] \sum b_i \overline{b_i} = 1 [/tex]
.

Now, this can be changed to:

[tex] < Q , P_{\sum_n b_n \psi_n} > = TR \ QP_{\sum_n b_n \psi_n} = [/tex]
[tex] \sum_n ( QP_{\sum_i b_i \psi_i} \psi_n , \psi_n ) = [/tex]
[tex] \sum_n ( Q \sum_j b_j \overline{b_n} \psi_j , \psi_n ) = [/tex]
[tex] \sum_n \overline{b_n} ( \sum_j b_j Q \psi_j , \psi_n ) = [/tex]
[tex] \sum_n \overline{b_n} \sum_j b_j ( Q \psi_j , \psi_n ) [/tex]

Compare that to:

[tex] <Q | M > = \sum_n a_n ( Q \psi_n , \psi_n ) [/tex]

If I set
[tex] Q = \sum c_n P_{\psi_n} [/tex]
, then
[tex] < Q | P_\psi > = \sum_n \overline{b_n} b_n c_n [/tex]

and

[tex] < Q | M > = \sum_n a_n c_n [/tex]

They are the same.

So for any Qs as [tex] \sum c_n P_{\psi_n} [/tex], we will have the same expectation value for the mixed state and the pure state.
 
  • #36
Eye,

You know sometimes this latex thing is strange.

My question is:

Any way, why do you say I is unbounded in a infinite dimensional space?

I mean,

[tex] ( I \psi , I \psi ) = ( \psi , \psi ) < 2 ( \psi , \psi ) [/tex]

whether the space is infinite or finite dimensional.

Thanks
 
  • #37
Any way, back to the issue between the "mixed" state and the "pure" states, further questions shall be:

1). Can we relax the conditions for the "pure" states and the observables A ( I wrote Q earlier , since this is a general observable not the "position", I think A is better)?

2). Even though they have the same expection value, shall they have different distribution such like < A | M > might have a multi-nodal distribution and < A | P_\psi > has a central normal distribution?
 
  • #38
Sammywu said:
Any way, why do you say I is unbounded in a infinite dimensional space?

I mean,

[tex] ( I \psi , I \psi ) = ( \psi , \psi ) < 2 ( \psi , \psi ) [/tex]
I forgot to mention the following:

Metatheorem: A necessary condition for I to be unbounded (in the ∞-dimensional case) is serious confusion. :surprise:


I'll have to fix that.
 
  • #39
Eye,

Now, let me take an example.

Assuming two free particles,

[tex] \psi_1 = \int \int \delta ( k - k_1 , w - w_1 ) e^{i(k(x-x_1)+w(t-t_1))} dk dw [/tex]

and

[tex] \psi_2 = \int \int \delta ( k - k_2 , w - w_2 ) e^{i(k(x-x_2)+w(t-t_2))} dk dw [/tex]

represent their wave functions; their states shall be
[tex] P_{\psi_1} \\ and \\ P_{\psi_2} [/tex].

In combination, this mixed state [tex] M = 1/2 ( P_{\psi_1} + P_{\psi_2} ) [/tex] shall represent their combined state, or the ensemble.

Correct?
 
  • #40
I see you are still using that notation where P is not a projector and its subscript isn't a unit vector:

[tex] P_{\sum_n a_n \psi_n} [/tex] .

According to that notation, is it not true that the preceding expression equals

[tex] \sum_n a_n P_{\psi_n } [/tex] ?

So, why don't you just use the second one?

-------------------------------

The step below has an error:

[tex] \sum_n ( QP_{\sum_i a_i \psi_i} \psi_n , \psi_n ) = \sum_n ( Q \sum_j a_j \overline{a_n} \psi_j , \psi_n ) [/tex]
I think this error results from what I just pointed out above, that you are treating this P as a projector when it is not! I also explained this same point in post #19:

[1] [tex] \sum_n a_n P_{e_n} = P_{\sum_n a_n e_n} [/tex]

The object on the left-hand-side is not (in general) a projector. That object has eigenvalues an , whereas a projector has eigenvalues 0 and / or 1.

---------------

[2] Pn(v) = (vn) n , for any vector v

----------------

I suggest you reserve the symbol "P", in the above type of context, only for a "projector" proper. Also, I suggest you invoke the rule that the "subscript" of P is always a "unit" vector. These two prescriptions would then disqualify the "legitimacy" of the right-hand-side of [1] on both counts.

At the same time, if you want to consider generalizations for which a relation like [1] holds, then use a symbol "R" (or whatever else) instead of "P". The generalization of [2] which gives a relation like [1] is then simply:

[2'] Ru(v) = [ v ∙ (u/|u|) ] u , for any vector v .

But what is the motivation for reserving a special "symbol" for this operation? Its description is "project the vector v into the direction of u and then multiply by the magnitude of u". The meaningful aspects of this operation are much better expressed by writing the corresponding operator as |u|P(u/|u|).
Do you follow what I am saying?

------

All of this means there is no second condition ∑nan2=1. (In which case, your post #35 no longer stands (I think :confused: ).)
 
Last edited:
  • #41
Eye,

Yes. I followed what you say. The correct defintion of projector here has a necessary condition in that the norm of the ket needs to be one or in other words it's a normal vector.

I correct that in the next post. [tex] \sum b_i \overline{b_i} = 1 [/tex] gurantees that [tex] \psi [/tex] is a normal vector.

Thanks
 
  • #42
Eye,

You are right.

I shall have used the second one to differentiate it from P.

OK.
 
  • #43
Sammywu said:
Any way, back to the issue between the "mixed" state and the "pure" states, further questions shall be:

1). Can we relax the conditions for the "pure" states and the observables A ( I wrote Q earlier , since this is a general observable not the "position", I think A is better)?

2). Even though they have the same expection value, shall they have different distribution such like < A | M > might have a multi-nodal distribution and < A | P_\psi > has a central normal distribution?
In 1): relax how? ... or is that what 2) is explaining?

I don't understand 2).
 
  • #44
Eye,

1) Relax.. means I could even find simpler conditions such as maybe all vectors in [tex] H_M [/tex] can be treated and found a role in this issue. Or maybe A as any function of [tex] P_{\psi_n} [/tex] can satisfy similar property.

My guess to the question is probably NO. The conditions I found pretty describe the situations of indistinguishable "mixed" and "pure" states.

2). What I am saying is even though their expectation values are the same. The statistics will show us two different distributions. So now it's time to investigate their probability decomposition of AM and [tex] AP_\psi [/tex].

Thoughts came out from me sometimes just are not described immediately in correct math. language but a pragmatic thought to begin with. So, I wrote distribution of < A | M >, things like that.

Multi-nodal distribution is my words for the distribution that you see multiple distingushiable points like two clear different frequency of light pulse will leave two lines in the light spectrum.

Central normal distribution is my words to emphasize that a normal probability distribution has a central node with an inverse exponetial shape of distribution.

I am sorry I do not remember what are correct math. terms for them.

Regards
 
  • #45
The example I show is two particles shot rightward at different times and at time T we shine rays of lights from top downward. Now will we predict from this math. the light detector at the bottom will show two spots instead one spot.
 
  • #46
A pure state is represented by a unit vector |φ>, or equivalently, by a density operator ρ = |φ><φ|. In that case, ρ2 = ρ.

Suppose we are unsure whether or not the state is |φ1> or |φ2>, but know enough to say that the state is |φi> with probability pi. Then the corresponding density operator is given by

ρ = p11><φ1| + p22><φ2| .

In that case ρ2 ≠ ρ, and the state is said to be mixed. Note that the two states |φ1> and |φ2> need not be orthogonal (however, if they are parallel (i.e. differ only by a phase factor), then we don't have mixed case but rather a pure case).

This is the real place to begin. And Dirac's notation is the superior way of representing these objects.

--------------------

Tell me, Sammy, have you learned the postulates of Quantum Mechanics in terms of simple, basic statements like the one below?

The probability of obtaining the result an in a measurement of the nondegenerate observable A on the system in the state |φ> is given by

P(an) = |<ψn|φ>|2 ,

where |ψn> is the eigenvector corresponding to the eigenvalue an.
 
Last edited:
  • #47
Eye,

I have seen that postulate but not fully convinced in that and tried to understand that and see whether my interpretation is correct.

So, I try to translate that here:

[tex] P(a_n) = ( \int \overline{\psi_n} \varphi )^2 [/tex]

Then
[tex] a_n P(a_n) = a_n ( \int \overline{\psi_n} \varphi )^2 = [/tex]
[tex] ( \int \overline{\psi_n} (a_n)^{1/2} \varphi )^2 [/tex]

Is there something wrong with my translation?

It seems that this is more likely.
[tex] P(a_n) = \int \overline{\psi_n} \varphi [/tex]
 
  • #48
I thought :

[tex] < A | M > = < \varphi | A | \varphi > = [/tex]
[tex] \int \overline{\varphi} (A) \varphi [/tex]

In order to lead us there.

If
[tex] P(a_n) = \int \overline{\psi_n} \varphi [/tex]

Then
[tex] \sum_n a_n P(a_n) = [/tex]
[tex] \int \sum_n ( a_n \overline{\psi_n} ) \varphi [/tex]

This might lead us there because [tex] a_n \psi_n = A \psi_n [/tex].
 
  • #49
Eye,

If I use Dirac notation,

[tex] < \varphi | A | \varphi > = [/tex]

[tex] < \varphi | A \sum_n | \psi_n > < \psi_n | \varphi> = [/tex]

[tex] \sum_n < \varphi | A | \psi_n > c_n = [/tex]
[tex] \sum_n \overline{c_n} c_n < \psi_n | A | \psi_n > = [/tex]
[tex] \sum_n \overline{c_n} c_n a_n = [/tex]
[tex] \sum_n | < \psi_n | \varphi > |^2 a_n [/tex]


Assuming [tex] < \psi_n | \varphi > = c_n [/tex]
, so
[tex] | \varphi > = \sum_n c_n | \psi_n > [/tex]
.
That do agree with your formula.

Does that look good to you?

But I do have some troubles to show that in an integral of Hermitian OP.
 
Last edited:
  • #50
Eye,

Back to what you said, the mixed state does not seem to apply to my proposed case.

I was confused about how we can use this mixed state. Now I have better idea but i still want to think further about it.

Any way, I thought the translation between the three notation is :

[tex] ( u , v ) = < v | u > = \int \overline{v} u [/tex]

Is this correct?
 
  • #51
Eye,

Is this where I did wrong?

I think I shall start with this.

[tex] P(a_n) = [/tex]
[tex] ( \int \overline{\psi_n(x)} \varphi (x) dx ) \overline{ ( \int \overline{\psi_n(y)} \varphi(y) dy ) } [/tex]

[tex] \sum_n a_n c_n \overline{c_n} = [/tex]
[tex] \sum_n a_n P(a_n) = [/tex]
[tex] \sum_n a_n \int \int \overline{\psi_n(x)} \varphi(x) \psi_n(y) \overline{\varphi(y)} dx dy = [/tex]
[tex] \sum_n \int \int \overline{\psi_n(x)} \varphi(x) A \psi_n(y) \overline{\varphi(y)} dx dy = [/tex]

I still need to see how this can be :

[tex] \int \overline{\varphi(x)} A \varphi(x) dx [/tex]



Thanks
 
Last edited:
  • #52
Eye,

Based on the two assumptions

[tex] \varphi = \sum_n c_n \psi_n [/tex]
and
[tex] \int \overline{\psi_i} \psi_j = \delta_{i,j} [/tex]
,

[tex] \int \overline{\varphi} A \varphi = [/tex]
[tex] \int \overline{\varphi} A \sum_n c_n \psi_n = [/tex]
[tex] \int \overline{\varphi} \sum_n c_n a_n \psi_n = [/tex]
[tex] \sum c_n a_n \int \overline{\varphi} \psi_n = [/tex]
[tex] \sum c_n a_n \overline{ \int \varphi overline{\psi_n} = [/tex]
[tex] \sum c_n a_n \overline{c_n} [/tex]

So, I verified all of these three approaches show the same expectation value.

Now, back to your question, why shall
[tex] P(a_n) = c_n \overline{c_n} [/tex]
?

Any good argument about it?
 
  • #53
Eye,

Now I see where you are leading.

The probability decomposition for AM will be

[tex] P(a) = [/tex]
[tex] \sum_{a_n<=a} P(a_n) = [/tex]
[tex] \sum_{a_n<=a} a_n c_n \overline{c_n} [/tex]
.

Is it?
 
  • #54
Now, if I use my original model of
[tex] M = 1/2 ( P_{\psi_1} + P_{\psi_2} ) [/tex]
or
[tex] M = 1/2 | \psi_1> < \psi_1 | + 1/2 | \psi_2> < \psi_2 | [/tex]
even though
[tex] < \psi_2 | \psi_1 > not = 0 [/tex]
and let
[tex] P_{XM}(a) [/tex]
be its probability decomposition,
this does seem to lead to a two spots measurement.

A trouble I might need to take care of is the two wavefunctions are not orthogonal.

While I recall that a multiple particle model in a book using a model
[tex] \psi = \psi_1 \psi_2 [/tex]
, I wonder how these two different models will turn out in this case.
 
  • #55
Any way, back to the basic postulate, it does now seem very reasonable for probability for eigenvalue

[tex] P(a_n) = < \psi_n | \varphi > < \varphi | \psi_n > = ( \psi_n , \varphi) ( \varphi , \psi_n ) = [/tex]
[tex] \int \overline{\psi_n} \varphi \int \psi_n \overline{\varphi} [/tex]
 
  • #56
If I use Dirac notation,

[tex] < \varphi | A | \varphi > = [/tex]

[tex] \sum_n | < \psi_n | \varphi > |^2 a_n [/tex]

Assuming [tex] < \psi_n | \varphi > = c_n [/tex]
, so
[tex] | \varphi > = \sum_n c_n | \psi_n > [/tex]

Does that look good to you?
It is correct.
_________

Any way, I thought the translation between the three notation is :

[tex] ( u , v ) = < v | u > = \int \overline{v} u [/tex]

Is this correct?
Looks fine.
_________
_________

... back to the basic postulate, it does now seem very reasonable for probability for eigenvalue

[tex] P(a_n) = < \psi_n | \varphi > < \varphi | \psi_n > = ( \psi_n , \varphi) ( \varphi , \psi_n ) = [/tex]
[tex] \int \overline{\psi_n} \varphi \int \psi_n \overline{\varphi} [/tex]
The basic postulates are the "true" starting point of Quantum Mechanics. From those postulates, one can then build the more complex versions which talk about "mixed states" and "expectation values" of observables ("nondegenerate" and "degenerate" cases) ... like on page 37 of that eBook.

Try to find a book in which those basic postulates are written in clear, concise terms. Those postulates should talk about a nondegenerate observable with a discrete set of eigenvalues, where the quantum system is in a pure state. There should also be a statement concerning the time evolution of a pure state in terms of the Schrödinger equation.

Once you understand and accept those basic postulates, then from them you will be able to derive - in a most transparent way - all of the more complex versions.

From that perspective, everything will be much clearer. I am quite sure of this.
 
  • #57
Eye,

I think I have a book that shall have a discussion of that postulate. I shall have read it but just going over it without too much deeper thought.

I always wanted to try thinking things in my logical reasoning and then compare that to what is there any way. So I tried to analyze what it is here.

What I have here is a state [tex] \varphi [/tex] representing a probability and all I know is
[tex] ( \varphi , \varphi ) = [/tex]
[tex] < \varphi , \varphi > = [/tex]
[tex] \int \overline{\varphi} \varphi = 1 [/tex]
.
Now with an observable A, I might have different observated value [tex] a_n [/tex] and I can associate with each one of them with a state [tex] \psi_n [/tex].

By the assumption of { [tex] \psi_n [/tex] } being a orthonornal basis,

[tex] \varphi = \sum_i c_i \psi_n [/tex]
.

The total probability as unity shall be decomposed into components representing probability for each [tex] a_n [/tex] .
[tex] 1 = (\varphi, \varphi ) = ( \sum_n c_n \psi_n , \varphi ) = \sum_n c_n ( \psi_n , \varphi ) [/tex]
.

Since this decomposition of unity has a coresponding number to each [tex] a_n [/tex], I can assume that the probaility for each a_n will be
[tex] c_n ( \psi_n , \varphi ) = ( \varphi , \psi_n ) (\psi_n , \varphi ) [/tex]
.

While analyzing this, there seems to be a condition that shall be in effect, i. e all of different possible outcomes as [tex] a_n [/tex] need to be independent to each other and that's what orthogonality of eigenfunctions guranteed.

If the a_n are not independent to each other, i.e. { [tex] \psi_n [/tex] } is not orthonormal, then there will be some interference can be derived here.

Any way, this is my own way to try to comprehend this postulate, but it does seem to make it clearer how this postulate was formulated and also indicate how the wavefunction interference might have come into play.
 
Last edited:
  • #58
Another important fact I noticed is that [tex] c \psi_n [/tex] is also an eigenfunction and regarded the same as [tex] \psi_n [/tex] as logon as | c | = 1; this is related to what is stationary state.

So [tex] \varphi [/tex] can be decomposed into a way such that all c_i are real. Still [tex] \sum_i c_i [/tex] does not equal to one but [tex] \sum_i c_i^2 = 1 [/tex] so that it will show me why [/tex] c_i^2 [/tex] not [tex] c_i [/tex] is the probability.

My intuition is c_i shall be the probability but not c_i^2, this makes it clearer to me why my intuition is wrong.
 
Last edited:
  • #59
I think I have a book that shall have a discussion of that postulate. I shall have read it but just going over it without too much deeper thought.

I always wanted to try thinking things in my logical reasoning and then compare that to what is there any way. So I tried to analyze what it is here.

What I have here is a state [tex] \varphi [/tex] representing a probability and all I know is
[tex] ( \varphi , \varphi ) = [/tex]
[tex] < \varphi , \varphi > = [/tex]
[tex] \int \overline{\varphi} \varphi = 1 [/tex]
.
Now with an observable A, I might have different observated value [tex] a_n [/tex] and I can associate with each one of them with a state [tex] \psi_n [/tex].

By the assumption of { [tex] \psi_n [/tex] } being a orthonornal basis,

[tex] \varphi = \sum_i c_i \psi_n [/tex]
.

The total probability as unity shall be decomposed into components representing probability for each [tex] a_n [/tex] .
[tex] 1 = (\varphi, \varphi ) = ( \sum_n c_n \psi_n , \varphi ) = \sum_n c_n ( \psi_n , \varphi ) [/tex]
.

Another way to look into this:

An abstract state can be represented by different orthonormal basis. By an unitary transformtion, we can translate a state to different representation in another basis, but its norm remians as one. So the square of its coefficiens can be conviniently used for probability representation. The basis is used as the "coordinate" of the measurable.

For example, Fourier transformation as an unitary transformation can transform a "position" represtation to a "momentum" representation.

An operator as an obserable can be only measured as certain real values; these are the eigenvalues and they work just like the "coordinate" of the measurement. Only certain probability distribution ( i.e. states ) can be exacted to one of these real values: they are the eigenfunctions and pure states.

For example, the eigenfunction of "position" observable of x_0 can be seen as the [tex] \delta ( x - x_0) [/tex] .

When measuring other states, only eigenvalues will appear, but since it has non-zero coefficients on different eigenvalues so it will show a dstribution among these eigen values.

An degenerate observable has more than two orthogonal functions measured the same value, so its probability measured for this value shall be the sum of the square of their ciefficients.
 
  • #60
Sammywu said:
... there seems to be a condition that shall be in effect, i. e all of different possible outcomes as [tex] a_n [/tex] need to be independent to each other and that's what orthogonality of eigenfunctions guranteed.
I am not quite sure what you mean by "independent of each other". If the |ψn> are not all mutually orthogonal, then the whole postulate 'falls apart'! ... we will no longer have ∑nP(an) = 1.

-----

Another important fact I noticed is that [tex] c \psi_n [/tex] is also an eigenfunction and regarded the same as [tex] \psi_n [/tex] as logon as | c | = 1; this is related to what is stationary state.
Yes, there is a connection.

------------------------------
------------------------------

Sammy, looking at all of the different points you are making, I get the sense that it might be instructive for us to go through each one of the postulates in a clear and concise way, one by one, step by step, ... . What do you say?
 

Similar threads

  • · Replies 19 ·
Replies
19
Views
4K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 13 ·
Replies
13
Views
2K
  • · Replies 3 ·
Replies
3
Views
6K