Understanding Projectors in Quantum Mechanics: A Mathematical Approach

In summary: Sum_n P_{e_n} = P_{\sum_n e_n} ...This is where things get a little fuzzy. It looks like you are saying that for every pair (\psi, \psi_n) in H,Sum_n P_{e_n} = P_{\sum_n e_n}
  • #36
Eye,

You know sometimes this latex thing is strange.

My question is:

Any way, why do you say I is unbounded in a infinite dimensional space?

I mean,

[tex] ( I \psi , I \psi ) = ( \psi , \psi ) < 2 ( \psi , \psi ) [/tex]

whether the space is infinite or finite dimensional.

Thanks
 
Physics news on Phys.org
  • #37
Any way, back to the issue between the "mixed" state and the "pure" states, further questions shall be:

1). Can we relax the conditions for the "pure" states and the observables A ( I wrote Q earlier , since this is a general observable not the "position", I think A is better)?

2). Even though they have the same expection value, shall they have different distribution such like < A | M > might have a multi-nodal distribution and < A | P_\psi > has a central normal distribution?
 
  • #38
Sammywu said:
Any way, why do you say I is unbounded in a infinite dimensional space?

I mean,

[tex] ( I \psi , I \psi ) = ( \psi , \psi ) < 2 ( \psi , \psi ) [/tex]
I forgot to mention the following:

Metatheorem: A necessary condition for I to be unbounded (in the ∞-dimensional case) is serious confusion. :surprise:


I'll have to fix that.
 
  • #39
Eye,

Now, let me take an example.

Assuming two free particles,

[tex] \psi_1 = \int \int \delta ( k - k_1 , w - w_1 ) e^{i(k(x-x_1)+w(t-t_1))} dk dw [/tex]

and

[tex] \psi_2 = \int \int \delta ( k - k_2 , w - w_2 ) e^{i(k(x-x_2)+w(t-t_2))} dk dw [/tex]

represent their wave functions; their states shall be
[tex] P_{\psi_1} \\ and \\ P_{\psi_2} [/tex].

In combination, this mixed state [tex] M = 1/2 ( P_{\psi_1} + P_{\psi_2} ) [/tex] shall represent their combined state, or the ensemble.

Correct?
 
  • #40
I see you are still using that notation where P is not a projector and its subscript isn't a unit vector:

[tex] P_{\sum_n a_n \psi_n} [/tex] .

According to that notation, is it not true that the preceding expression equals

[tex] \sum_n a_n P_{\psi_n } [/tex] ?

So, why don't you just use the second one?

-------------------------------

The step below has an error:

[tex] \sum_n ( QP_{\sum_i a_i \psi_i} \psi_n , \psi_n ) = \sum_n ( Q \sum_j a_j \overline{a_n} \psi_j , \psi_n ) [/tex]
I think this error results from what I just pointed out above, that you are treating this P as a projector when it is not! I also explained this same point in post #19:

[1] [tex] \sum_n a_n P_{e_n} = P_{\sum_n a_n e_n} [/tex]

The object on the left-hand-side is not (in general) a projector. That object has eigenvalues an , whereas a projector has eigenvalues 0 and / or 1.

---------------

[2] Pn(v) = (vn) n , for any vector v

----------------

I suggest you reserve the symbol "P", in the above type of context, only for a "projector" proper. Also, I suggest you invoke the rule that the "subscript" of P is always a "unit" vector. These two prescriptions would then disqualify the "legitimacy" of the right-hand-side of [1] on both counts.

At the same time, if you want to consider generalizations for which a relation like [1] holds, then use a symbol "R" (or whatever else) instead of "P". The generalization of [2] which gives a relation like [1] is then simply:

[2'] Ru(v) = [ v ∙ (u/|u|) ] u , for any vector v .

But what is the motivation for reserving a special "symbol" for this operation? Its description is "project the vector v into the direction of u and then multiply by the magnitude of u". The meaningful aspects of this operation are much better expressed by writing the corresponding operator as |u|P(u/|u|).
Do you follow what I am saying?

------

All of this means there is no second condition ∑nan2=1. (In which case, your post #35 no longer stands (I think :confused: ).)
 
Last edited:
  • #41
Eye,

Yes. I followed what you say. The correct defintion of projector here has a necessary condition in that the norm of the ket needs to be one or in other words it's a normal vector.

I correct that in the next post. [tex] \sum b_i \overline{b_i} = 1 [/tex] gurantees that [tex] \psi [/tex] is a normal vector.

Thanks
 
  • #42
Eye,

You are right.

I shall have used the second one to differentiate it from P.

OK.
 
  • #43
Sammywu said:
Any way, back to the issue between the "mixed" state and the "pure" states, further questions shall be:

1). Can we relax the conditions for the "pure" states and the observables A ( I wrote Q earlier , since this is a general observable not the "position", I think A is better)?

2). Even though they have the same expection value, shall they have different distribution such like < A | M > might have a multi-nodal distribution and < A | P_\psi > has a central normal distribution?
In 1): relax how? ... or is that what 2) is explaining?

I don't understand 2).
 
  • #44
Eye,

1) Relax.. means I could even find simpler conditions such as maybe all vectors in [tex] H_M [/tex] can be treated and found a role in this issue. Or maybe A as any function of [tex] P_{\psi_n} [/tex] can satisfy similar property.

My guess to the question is probably NO. The conditions I found pretty describe the situations of indistinguishable "mixed" and "pure" states.

2). What I am saying is even though their expectation values are the same. The statistics will show us two different distributions. So now it's time to investigate their probability decomposition of AM and [tex] AP_\psi [/tex].

Thoughts came out from me sometimes just are not described immediately in correct math. language but a pragmatic thought to begin with. So, I wrote distribution of < A | M >, things like that.

Multi-nodal distribution is my words for the distribution that you see multiple distingushiable points like two clear different frequency of light pulse will leave two lines in the light spectrum.

Central normal distribution is my words to emphasize that a normal probability distribution has a central node with an inverse exponetial shape of distribution.

I am sorry I do not remember what are correct math. terms for them.

Regards
 
  • #45
The example I show is two particles shot rightward at different times and at time T we shine rays of lights from top downward. Now will we predict from this math. the light detector at the bottom will show two spots instead one spot.
 
  • #46
A pure state is represented by a unit vector |φ>, or equivalently, by a density operator ρ = |φ><φ|. In that case, ρ2 = ρ.

Suppose we are unsure whether or not the state is |φ1> or |φ2>, but know enough to say that the state is |φi> with probability pi. Then the corresponding density operator is given by

ρ = p11><φ1| + p22><φ2| .

In that case ρ2 ≠ ρ, and the state is said to be mixed. Note that the two states |φ1> and |φ2> need not be orthogonal (however, if they are parallel (i.e. differ only by a phase factor), then we don't have mixed case but rather a pure case).

This is the real place to begin. And Dirac's notation is the superior way of representing these objects.

--------------------

Tell me, Sammy, have you learned the postulates of Quantum Mechanics in terms of simple, basic statements like the one below?

The probability of obtaining the result an in a measurement of the nondegenerate observable A on the system in the state |φ> is given by

P(an) = |<ψn|φ>|2 ,

where |ψn> is the eigenvector corresponding to the eigenvalue an.
 
Last edited:
  • #47
Eye,

I have seen that postulate but not fully convinced in that and tried to understand that and see whether my interpretation is correct.

So, I try to translate that here:

[tex] P(a_n) = ( \int \overline{\psi_n} \varphi )^2 [/tex]

Then
[tex] a_n P(a_n) = a_n ( \int \overline{\psi_n} \varphi )^2 = [/tex]
[tex] ( \int \overline{\psi_n} (a_n)^{1/2} \varphi )^2 [/tex]

Is there something wrong with my translation?

It seems that this is more likely.
[tex] P(a_n) = \int \overline{\psi_n} \varphi [/tex]
 
  • #48
I thought :

[tex] < A | M > = < \varphi | A | \varphi > = [/tex]
[tex] \int \overline{\varphi} (A) \varphi [/tex]

In order to lead us there.

If
[tex] P(a_n) = \int \overline{\psi_n} \varphi [/tex]

Then
[tex] \sum_n a_n P(a_n) = [/tex]
[tex] \int \sum_n ( a_n \overline{\psi_n} ) \varphi [/tex]

This might lead us there because [tex] a_n \psi_n = A \psi_n [/tex].
 
  • #49
Eye,

If I use Dirac notation,

[tex] < \varphi | A | \varphi > = [/tex]

[tex] < \varphi | A \sum_n | \psi_n > < \psi_n | \varphi> = [/tex]

[tex] \sum_n < \varphi | A | \psi_n > c_n = [/tex]
[tex] \sum_n \overline{c_n} c_n < \psi_n | A | \psi_n > = [/tex]
[tex] \sum_n \overline{c_n} c_n a_n = [/tex]
[tex] \sum_n | < \psi_n | \varphi > |^2 a_n [/tex]


Assuming [tex] < \psi_n | \varphi > = c_n [/tex]
, so
[tex] | \varphi > = \sum_n c_n | \psi_n > [/tex]
.
That do agree with your formula.

Does that look good to you?

But I do have some troubles to show that in an integral of Hermitian OP.
 
Last edited:
  • #50
Eye,

Back to what you said, the mixed state does not seem to apply to my proposed case.

I was confused about how we can use this mixed state. Now I have better idea but i still want to think further about it.

Any way, I thought the translation between the three notation is :

[tex] ( u , v ) = < v | u > = \int \overline{v} u [/tex]

Is this correct?
 
  • #51
Eye,

Is this where I did wrong?

I think I shall start with this.

[tex] P(a_n) = [/tex]
[tex] ( \int \overline{\psi_n(x)} \varphi (x) dx ) \overline{ ( \int \overline{\psi_n(y)} \varphi(y) dy ) } [/tex]

[tex] \sum_n a_n c_n \overline{c_n} = [/tex]
[tex] \sum_n a_n P(a_n) = [/tex]
[tex] \sum_n a_n \int \int \overline{\psi_n(x)} \varphi(x) \psi_n(y) \overline{\varphi(y)} dx dy = [/tex]
[tex] \sum_n \int \int \overline{\psi_n(x)} \varphi(x) A \psi_n(y) \overline{\varphi(y)} dx dy = [/tex]

I still need to see how this can be :

[tex] \int \overline{\varphi(x)} A \varphi(x) dx [/tex]



Thanks
 
Last edited:
  • #52
Eye,

Based on the two assumptions

[tex] \varphi = \sum_n c_n \psi_n [/tex]
and
[tex] \int \overline{\psi_i} \psi_j = \delta_{i,j} [/tex]
,

[tex] \int \overline{\varphi} A \varphi = [/tex]
[tex] \int \overline{\varphi} A \sum_n c_n \psi_n = [/tex]
[tex] \int \overline{\varphi} \sum_n c_n a_n \psi_n = [/tex]
[tex] \sum c_n a_n \int \overline{\varphi} \psi_n = [/tex]
[tex] \sum c_n a_n \overline{ \int \varphi overline{\psi_n} = [/tex]
[tex] \sum c_n a_n \overline{c_n} [/tex]

So, I verified all of these three approaches show the same expectation value.

Now, back to your question, why shall
[tex] P(a_n) = c_n \overline{c_n} [/tex]
?

Any good argument about it?
 
  • #53
Eye,

Now I see where you are leading.

The probability decomposition for AM will be

[tex] P(a) = [/tex]
[tex] \sum_{a_n<=a} P(a_n) = [/tex]
[tex] \sum_{a_n<=a} a_n c_n \overline{c_n} [/tex]
.

Is it?
 
  • #54
Now, if I use my original model of
[tex] M = 1/2 ( P_{\psi_1} + P_{\psi_2} ) [/tex]
or
[tex] M = 1/2 | \psi_1> < \psi_1 | + 1/2 | \psi_2> < \psi_2 | [/tex]
even though
[tex] < \psi_2 | \psi_1 > not = 0 [/tex]
and let
[tex] P_{XM}(a) [/tex]
be its probability decomposition,
this does seem to lead to a two spots measurement.

A trouble I might need to take care of is the two wavefunctions are not orthogonal.

While I recall that a multiple particle model in a book using a model
[tex] \psi = \psi_1 \psi_2 [/tex]
, I wonder how these two different models will turn out in this case.
 
  • #55
Any way, back to the basic postulate, it does now seem very reasonable for probability for eigenvalue

[tex] P(a_n) = < \psi_n | \varphi > < \varphi | \psi_n > = ( \psi_n , \varphi) ( \varphi , \psi_n ) = [/tex]
[tex] \int \overline{\psi_n} \varphi \int \psi_n \overline{\varphi} [/tex]
 
  • #56
If I use Dirac notation,

[tex] < \varphi | A | \varphi > = [/tex]

[tex] \sum_n | < \psi_n | \varphi > |^2 a_n [/tex]

Assuming [tex] < \psi_n | \varphi > = c_n [/tex]
, so
[tex] | \varphi > = \sum_n c_n | \psi_n > [/tex]

Does that look good to you?
It is correct.
_________

Any way, I thought the translation between the three notation is :

[tex] ( u , v ) = < v | u > = \int \overline{v} u [/tex]

Is this correct?
Looks fine.
_________
_________

... back to the basic postulate, it does now seem very reasonable for probability for eigenvalue

[tex] P(a_n) = < \psi_n | \varphi > < \varphi | \psi_n > = ( \psi_n , \varphi) ( \varphi , \psi_n ) = [/tex]
[tex] \int \overline{\psi_n} \varphi \int \psi_n \overline{\varphi} [/tex]
The basic postulates are the "true" starting point of Quantum Mechanics. From those postulates, one can then build the more complex versions which talk about "mixed states" and "expectation values" of observables ("nondegenerate" and "degenerate" cases) ... like on page 37 of that eBook.

Try to find a book in which those basic postulates are written in clear, concise terms. Those postulates should talk about a nondegenerate observable with a discrete set of eigenvalues, where the quantum system is in a pure state. There should also be a statement concerning the time evolution of a pure state in terms of the Schrödinger equation.

Once you understand and accept those basic postulates, then from them you will be able to derive - in a most transparent way - all of the more complex versions.

From that perspective, everything will be much clearer. I am quite sure of this.
 
  • #57
Eye,

I think I have a book that shall have a discussion of that postulate. I shall have read it but just going over it without too much deeper thought.

I always wanted to try thinking things in my logical reasoning and then compare that to what is there any way. So I tried to analyze what it is here.

What I have here is a state [tex] \varphi [/tex] representing a probability and all I know is
[tex] ( \varphi , \varphi ) = [/tex]
[tex] < \varphi , \varphi > = [/tex]
[tex] \int \overline{\varphi} \varphi = 1 [/tex]
.
Now with an observable A, I might have different observated value [tex] a_n [/tex] and I can associate with each one of them with a state [tex] \psi_n [/tex].

By the assumption of { [tex] \psi_n [/tex] } being a orthonornal basis,

[tex] \varphi = \sum_i c_i \psi_n [/tex]
.

The total probability as unity shall be decomposed into components representing probability for each [tex] a_n [/tex] .
[tex] 1 = (\varphi, \varphi ) = ( \sum_n c_n \psi_n , \varphi ) = \sum_n c_n ( \psi_n , \varphi ) [/tex]
.

Since this decomposition of unity has a coresponding number to each [tex] a_n [/tex], I can assume that the probaility for each a_n will be
[tex] c_n ( \psi_n , \varphi ) = ( \varphi , \psi_n ) (\psi_n , \varphi ) [/tex]
.

While analyzing this, there seems to be a condition that shall be in effect, i. e all of different possible outcomes as [tex] a_n [/tex] need to be independent to each other and that's what orthogonality of eigenfunctions guranteed.

If the a_n are not independent to each other, i.e. { [tex] \psi_n [/tex] } is not orthonormal, then there will be some interference can be derived here.

Any way, this is my own way to try to comprehend this postulate, but it does seem to make it clearer how this postulate was formulated and also indicate how the wavefunction interference might have come into play.
 
Last edited:
  • #58
Another important fact I noticed is that [tex] c \psi_n [/tex] is also an eigenfunction and regarded the same as [tex] \psi_n [/tex] as logon as | c | = 1; this is related to what is stationary state.

So [tex] \varphi [/tex] can be decomposed into a way such that all c_i are real. Still [tex] \sum_i c_i [/tex] does not equal to one but [tex] \sum_i c_i^2 = 1 [/tex] so that it will show me why [/tex] c_i^2 [/tex] not [tex] c_i [/tex] is the probability.

My intuition is c_i shall be the probability but not c_i^2, this makes it clearer to me why my intuition is wrong.
 
Last edited:
  • #59
I think I have a book that shall have a discussion of that postulate. I shall have read it but just going over it without too much deeper thought.

I always wanted to try thinking things in my logical reasoning and then compare that to what is there any way. So I tried to analyze what it is here.

What I have here is a state [tex] \varphi [/tex] representing a probability and all I know is
[tex] ( \varphi , \varphi ) = [/tex]
[tex] < \varphi , \varphi > = [/tex]
[tex] \int \overline{\varphi} \varphi = 1 [/tex]
.
Now with an observable A, I might have different observated value [tex] a_n [/tex] and I can associate with each one of them with a state [tex] \psi_n [/tex].

By the assumption of { [tex] \psi_n [/tex] } being a orthonornal basis,

[tex] \varphi = \sum_i c_i \psi_n [/tex]
.

The total probability as unity shall be decomposed into components representing probability for each [tex] a_n [/tex] .
[tex] 1 = (\varphi, \varphi ) = ( \sum_n c_n \psi_n , \varphi ) = \sum_n c_n ( \psi_n , \varphi ) [/tex]
.

Another way to look into this:

An abstract state can be represented by different orthonormal basis. By an unitary transformtion, we can translate a state to different representation in another basis, but its norm remians as one. So the square of its coefficiens can be conviniently used for probability representation. The basis is used as the "coordinate" of the measurable.

For example, Fourier transformation as an unitary transformation can transform a "position" represtation to a "momentum" representation.

An operator as an obserable can be only measured as certain real values; these are the eigenvalues and they work just like the "coordinate" of the measurement. Only certain probability distribution ( i.e. states ) can be exacted to one of these real values: they are the eigenfunctions and pure states.

For example, the eigenfunction of "position" observable of x_0 can be seen as the [tex] \delta ( x - x_0) [/tex] .

When measuring other states, only eigenvalues will appear, but since it has non-zero coefficients on different eigenvalues so it will show a dstribution among these eigen values.

An degenerate observable has more than two orthogonal functions measured the same value, so its probability measured for this value shall be the sum of the square of their ciefficients.
 
  • #60
Sammywu said:
... there seems to be a condition that shall be in effect, i. e all of different possible outcomes as [tex] a_n [/tex] need to be independent to each other and that's what orthogonality of eigenfunctions guranteed.
I am not quite sure what you mean by "independent of each other". If the |ψn> are not all mutually orthogonal, then the whole postulate 'falls apart'! ... we will no longer have ∑nP(an) = 1.

-----

Another important fact I noticed is that [tex] c \psi_n [/tex] is also an eigenfunction and regarded the same as [tex] \psi_n [/tex] as logon as | c | = 1; this is related to what is stationary state.
Yes, there is a connection.

------------------------------
------------------------------

Sammy, looking at all of the different points you are making, I get the sense that it might be instructive for us to go through each one of the postulates in a clear and concise way, one by one, step by step, ... . What do you say?
 
  • #61
Eye,

What I was saying about independency is that it came to my mind that interference of two wave functions will be zero when they are orthogonal. Their interference seems to be represented by their inner product.

Of course, I know the necessary condition in here is that these are all orthonormal basises. I am saying is we can ignore their interference because they are orthogonal, but it would be interesting to check the relationship between interference and the inner product of two arbitrary wave functions.

I think your suggestion of going over all postulations is good.

Thanks
 
  • #62
Sammywu said:
What I was saying about independency is that it came to my mind that interference of two wave functions will be zero when they are orthogonal. Their interference seems to be represented by their inner product.
Yes, I see.

If

|φ> = c11> + c22> ,

then

<φ|φ> = |c1|211> + |c2|222> + 2 Re{c1*c212>} .

The last term is the "interference" term, and it vanishes if <φ12> = 0 .

------------------

Is the above what you meant?

... it would be interesting to check the relationship between interference and the inner product of two arbitrary wave functions.
------------------
------------------

I think your suggestion of going over all postulations is good.
I will post something soon.
 
  • #63
Two Beginning Postulates of QM

ATTENTION: Anyone following this thread ... if you find any points (or 'near' points) of error on my part, please do point them out.
___________

P0: To a quantum system S there corresponds an associated Hilbert space HS.

P1: A pure state of S is represented by a ray (i.e. a one-dimensional subspace) of HS.
___________

Notes:

N.0.1) Regarding postulate P0, in certain contexts it is possible to associate a Hilbert space with a particular dynamical 'aspect' of the quantum system (e.g. a Hilbert space corresponding to "spin", decoupled from, say, "position").

N.1.1) In postulate P1, a "ray" is understood to represent the "(pure) state" of a single quantum system. Some physicists prefer to let those terms designate an ensemble of "identically prepared" quantum systems. Such a distinction becomes relevant only in cases where one considers possible interpretations of the theory.

N.1.2) A ray is determined by anyone of its (non-zero) vectors. Our convention is to use a "normalized" (i.e. to unity) ket |ψ> to designate the corresponding ray, and hence, the corresponding pure state. This means that two normalized kets |ψ> and |ψ'> which differ by only a phase factor (i.e. |ψ'> = α|ψ>, where |α| = 1) will represent the same "ray", and hence, the same "state".
___________

Exercise:

E.1.1) What, if anything, is wrong with the following?

Suppose that the kets |φ1> and |φ1'> represent the same state. Then, the kets |ψ> and |ψ'> given below will also represent the same state:

|ψ> = c11> + c22> ,

|ψ'> = c11'> + c22> .
___________
 
Last edited:
  • #64
Eye,

About the interference of wave functions: Yes. You got what I meant. I will have to ruminate over your simple answer, though.

Answer to the exercise:

If
[tex] | \psi \prime > = a | \psi > [/tex]
[tex] | \varphi_1 \prime > = a_1 | \varphi_1 > [/tex]
[tex] | \varphi_2 \prime > = a_1 | \varphi_2 > [/tex]
then
[tex] a c_1 | \varphi_1 > + a c_2 | \varphi_2 > = [/tex]
[tex] a_1 c_1 | \varphi_1 > + a_2 c_2 | \varphi_2 > [/tex]
.

One solution to it is:
[tex] a = a_1 = a_2 [/tex]
i. e. they are mutiplied by the same phase factor.

If
[tex] a not = a_1 [/tex]
, then
[tex] | \varphi_1 > = ( a_2 -a ) c_2 / ( a - a_1 ) c_1 | \varphi_2 > [/tex]
; i. e. [tex] | \varphi_1 > [/tex] and [tex] | \varphi_2 > [/tex] have to be the same state.

This shows that statement is in general incorrect except in the two solutions I show above.
 
  • #65
Actually, for the second solution of them being the same state, there are other troubles to take care, | (a_1 - a ) / (a - a_2 ) \ needs to be one.

It does not seem to be easy to get solution for this. I have to think about how to resolve it. The possibility is this solution will not even work.
 
  • #66
| ( a_1 - a ) c_1 / ( a_2 -a ) c_2 | =1 is equivalent to | a_1 - a | / | a_2 -a | = | c_2 | / | c_1 | ; knowing | a | = | a_1 | = | a_ 2 | = 1, we can view a, a_1 and a_2 as three normal vectors in the unit circle on the Complex plane; a shall be chosen as a point on the unit circle such that ratio of the cord length between a_1 and a over the cord length a_2 and a is | c_2 | / | c_1| .
 
Last edited:
  • #67
First off, you have attempted to answer a question slightly more general than the one I posed. In my question, there was no |φ2'>. But that doesn't really matter. ... We'll use your version of the question.

You start off well.
If
[tex] | \psi \prime > = a | \psi > [/tex]
[tex] | \varphi_1 \prime > = a_1 | \varphi_1 > [/tex]
[tex] | \varphi_2 \prime > = a_1 | \varphi_2 > [/tex]
then
[tex] a c_1 | \varphi_1 > + a c_2 | \varphi_2 > = [/tex]
[tex] a_1 c_1 | \varphi_1 > + a_2 c_2 | \varphi_2 > [/tex]
But your final conclusion suggests to me that the main point has been missed.
This shows that statement is in general incorrect except in the two solutions ...
Let's look at your second "solution". You write:
If
[tex] a not = a_1 [/tex]
, then
[tex] | \varphi_1 > = ( a_2 -a ) c_2 / ( a - a_1 ) c_1 | \varphi_2 > [/tex]
; i. e. [tex] | \varphi_1 > [/tex] and [tex] | \varphi_2 > [/tex] have to be the same state.
In this case, you are right. If |φ1> and |φ2> themselves represent the same state, then so too will |ψ> and |ψ'>.

Now, let's look at your first "solution".
One solution to it is:
[tex] a = a_1 = a_2 [/tex]
i. e. they are mutiplied by the same phase factor.
Can we say that the statement is correct in this case? For the statement to be correct, we must have a1 and a2 as arbitrary free parameters, except for the constraint |a1| = |a2| = 1; but this solution produces an additional constraint over and above that.

... So, your conclusion should have been:

The statement is incorrect in all cases, except when |φ1> and |φ2> represent the same state.

------------------------

NOW ...
Just to make sure that the main point hasn't been lost in all of this abstraction, let's look at a concrete example.

Given the state

c11> + c22> ,

then

i(c11> + c22>)

represents the same state,

but

ic11> + c22>

does not (unless, of course, |φ1> and |φ2> represent the same state).

In the first case, we have inserted a "global phase-factor", and that is OK. In the second case, however, we have inserted a "relative phase-factor", and that is not OK.

------------------------

Finally, for the sake of completeness, I offer a solution of my own along the same lines as the one you gave above.
____

First, we assume:

(i) c1, c2 ≠ 0 ,

and

(ii) |φ1> and |φ2> are linearly independent .

For, otherwise, it follows trivially that |ψ> and |ψ'> will define the same ray (i.e. represent the same state).


Next, by considerations similar to yours above, we note that

|ψ> and |ψ'> define the same ray if, and only if

[1] c1(a - a1)|φ1> + c2(a - a2)|φ2> = 0 ,

where (and this is the important part!) the parameters a1 and a2 are completely arbitrary except for the constraint |a1| = |a2| = 1.

We now reach our conclusion. We say: but from assumptions (i) and (ii) we see that relation [1] holds iff a = a1 = a2, implying that a1 and a2 cannot be completely arbitrary (except for the constraint) as required; therefore, |ψ> and |ψ'> do not define the same ray.

:smile:
------------------------
 
Last edited:
  • #68
Eye,

I am sorry. I did not notice there is no prime on the second \varphi_2.

I agree to your answer.

One part of my calculation was wrong. Somehow | c_1 | / | c_2| = 1 slipped into my mind, which is not true, so I corrected my answer in case somebody else were reading this post. Any way, that was showing even if they are all "same" state, "a" need to be chosen in a correct math. way.
 
Last edited:
  • #69
Another Postulate

P2: To a physical quantity A measurable on (the quantum system) S, there corresponds a self-adjoint linear operator A acting in HS. Such an operator is said to be an "observable".
___________

Notes:

N.2.1) From "The Spectral Theorem" for self-adjoint operators, it follows that:

(a) A has real eigenvalues;

(b) the eigenvectors corresponding to distinct eigenvalues are orthogonal;

(c) the eigenvectors of A are complete (i.e. they span HS).

The set of eigenvalues of A is called the "spectrum" of A. If A has a continuous spectrum, then the eigenvectors of A are said to be "generalized" and A is said to satisfy a "generalized" eigenvalue equation.

N.2.2) We now enumerate three special cases for A:

(1) A has a discrete, nondegenerate spectrum; then

A = ∑n an|u><un| .

(2) A has a discrete (possibly degenerate) spectrum; then

A = ∑n anPn .

(3) A has a continuous, nondegenerate spectrum; then

A = ∫a |a><a| da .

In each case, the RHS of each of the above relations is referred to as the "spectral decomposition" for A.

Case (1) is a particularization of case (2) (with Pn = |un><un|). For the general case of (2), the an are the eigenvalues of A and the Pn are the corresponding eigenprojectors.

Examples of case (3) are any of the components (Qj or Pk) of the position or momentum observables.
___________

Exercises:

E.2.1) From N.2.1) (b) (i.e. the eigenvectors corresponding to distinct eigenvalues are orthogonal), show that for a spectral decomposition of the type in N.2.2) (2), i.e.

A = ∑n anPn ,

it follows that

PjPk = δjkPk .

From N.2.1) (c) (i.e. the eigenvectors of A are complete), show that

nPn = I .

E.2.2) Use, as an example, the observable Q (for the position of a spinless particle moving in 1-dimension) to explain why the eigenkets |q> are said to be "generalized" and, therefore, why Q is said to satisfy a "generalized" eigenvalue equation.
___________
 
  • #70
Eye,

I).
I think you have a typo here. Did you miss a subscriptor n here in this paragraph.

N.2.2) We now enumerate three special cases for A:

(1) A has a discrete, nondegenerate spectrum; then

A = ∑n an|un><un| .

II)
Also, For acontinuous spectrum, why are the eigenvectors called " generalized"? Does this have anythings to do with the fact that most likely they will be "generalized" function such as dirac-delta function isntead of regular functions?

III). Answer to the Exercise. It's actually a little difficult because you did not define eigenprojectors.

So, I just gave a guess on this:

[tex] P_n ( \sum_i c_i u_n_i + b u_\bot ) = \sum_i c_i u_n_i [/tex]
where
[tex] u_n_i [/tex] are eigenvectors of [tex] a_n [/tex] and [tex] u_\bot \bot [/tex] all [tex] u_n [/tex] .

Under that assumption,
[tex] P_n^2 ( \sum_i c_i u_n_i + b u_\bot ) = P_n ( \sum_i c_i u_n_i ) =[/tex]
[tex] \sum_i c_i u_n_i [/tex]

so, [tex] P_n^2 = P_n [/tex] .

If i does not equal to j, then
[tex] P_i P_J ( \sum_k c_k u_i_k + \sum_l c_l u_j_l + b u_\bot ) = [/tex]
[tex] P_i ( \sum_l c_l u_j_l ) = 0 [/tex]
; so [tex] P_i P_j = 0 [/tex] .

That shall take care of the first part of E.2.1).

The question could be asked is how I can prove all u can be decomposed to
[tex] \sum_k c_k u_i_k + \sum_l c_l u_j_l + b u_\bot [/tex]
.

I think setting
[tex] c_k = < u_i_k | u > [/tex]
[tex] c_l = < u_j_l | u > [/tex]
and
[tex] b u_\bot = u - \sum_k c_k u_i_k - \sum_l c_l u_j_l [/tex]
could take care of that.

I might verify this part later.

I am still working on the other two exercises.
 

Similar threads

  • Quantum Physics
Replies
19
Views
3K
Replies
5
Views
1K
  • Introductory Physics Homework Help
Replies
13
Views
2K
  • Advanced Physics Homework Help
Replies
3
Views
5K
Back
Top