Is Every Hermitian Operator Observable?

  • Thread starter Thread starter sweet springs
  • Start date Start date
  • Tags Tags
    Observable
Click For Summary
Not every Hermitian operator is an observable, as demonstrated by examples like the identity operator, which, while Hermitian, does not provide meaningful measurement information due to its constant eigenvalue. The number operator is a valid observable since it yields measurable quantum numbers in specific states. Observables are defined operationally, meaning they must correspond to measurable quantities through devices, not just mathematically as Hermitian operators. The discussion highlights the distinction between mathematical definitions of observables and their practical measurement implications in quantum mechanics. Understanding these nuances is essential for grasping the relationship between Hermitian operators and observables in quantum theory.
  • #31
Hi.

Fredrik said:
Some of the symbols you're typing don't display properly for me, on either of my two computers.

Excuse me. I will restate my questions.

Definition:
OBSERVABLE is operator whose eigenvectors form a complete set (by Dirac).
OBSERVABLE is operator whose eigenspaces contain all the maximal orthonormal set, i.e. basis(Thanks to Fredrik).
Both the definition are equivalent.

Question:Are the following operators OBSERVABLE?
-Identity operator
-Null operator
-Projection to a subspace e.g. |a1><a1| for A|an>=an|an> with eigenvalues { an| a1,a2,a3,...}

I want to know how to deal with "eigenspace with eigenvalue 0".

Regards.
 
Last edited:
Physics news on Phys.org
  • #32
strangerep said:
A "linear combination" in an arbitrary vector space can certainly be an infinite sum.

For an arbitrary vector space, what does "infinite sum" mean? There is only one topology, the Euclidean topology, that can be given to a finite-dimensional vector space, but an infinite-dimensional vector space can be given various topologies.
 
Last edited:
  • #33
sweet springs said:
Definition:
OBSERVABLE is operator whose eigenvectors form a complete set (by Dirac).
OBSERVABLE is operator whose eigenspaces contain all the maximal orthonormal set, i.e. basis(Thanks to Fredrik).
Both the definition are equivalent.

Question:Are the following operators OBSERVABLE?
-Identity operator
-Null operator
-Projection to a subspace e.g. |a1><a1| for A|an>=an|an> with eigenvalues { an| a1,a2,a3,...}

I want to know how to deal with "eigenspace with eigenvalue 0".
As I said, I'm not a fan of that definition, but given that definition, then all of those operators are observables. Recall that an eigenvector of a linear operator A is a non-zero vector x such that Ax=ax for some number a. Every non-zero vector is an eigenvector of the identity operator with eigenvalue 1. Every non-zero vector is an eigenvector of the null operator with eigenvalue 0. If P is a projection operator for a subspace V, then every non-zero member of V is an eigenvector of P with eigenvalue 1, and every non-zero vector that's orthogonal to all the vectors in V is an eigenvector of P with eigenvalue 0.

Eigenvalue 0 doesn't cause any additional complications at all, so it doesn't need to be handled separately.
 
  • #34
Fredrik said:
I don't know this stuff myself, but I get that the basic idea is to start with a C*-algebra, define a "state" as a positive linear functional on the C*-algebra of observables, and then invoke the appropriate mathematical theorems to prove that abelian C*-algebras give us classical theories and non-abelian C*-algebras give us quantum theories. (The C*-algebra is then isomorphic to the algebra of bounded self-adjoint operators on a complex separable
Hilbert space).
I just wanted to say something about this. An Abelian C*-algebra gives us a Probability theory, that is Kolmolgorov Probability. You may think of this as the kind of probability that results from our lack of knowledge. Non-Abelian gives us Quantum Theory. This is the mathematical insight of C*-algebras. Just like Non-Euclidean geometry is a generalisation of Euclidean geometry, Quantum Mechanics is a generalisation of Probability theory.

For instance the smallest and simplest Commutative C*-algebra is the Probability theory of a coin toss, 1/2 chance for tails, 1/2 chance for heads. The smallest Non-Commutative C*-algebra is the quantum theory of one particle with spin 1/2 and no other properties.

This actually provides a method of proving Bell's inequalities. If there was another theory underlying QM which gave definite values to things independant of context, then QM would simply be a result of our ignorance and the "randomness" would be due to a lack of knowledge. If this was the case the probability would be just usual probability, described by a commutative C*-algebra. However you can construct a set observables which must have correlations less than some value c in a commutative C*-algebra, but can have correlations exceeding c in Non-Commutative C*-algebras. Hence the predictions of QM are fundamentally different from such theories. Experiment supports the Non-Commutative C*-algebras and so QM is not just the result of our ignorance of an underlying theory.
By the way this proof does not assume locality.

Also the Non-Commutative C*-algebra approach provides what is in my opinion the best explanation of entanglement. As we all know, correlation is not causation. When we set up entangled particles and send them far apart to be measured we can easily see that the results are strongly correlated. However if we do the further statistical tests for causation then the correlations fail these tests. Hence the particles are not influencing each other. What then is the cause of these strange correlations? Simple, we are not dealing with old fashioned 19th/early 20th century probability. QM is a new theory of probability, one that allows stronger correlations than before. So it is not that the particles are causing effects on each other, but rather that they are more strongly correlated than is possible in probabilities that result from ignorance.

Fredrik said:
In this approach, it's not the case that every member of the C*-algebra corresponds to a measuring device, but I don't really have any more information on that. Perhaps someone can read those books and tell the rest of us. :smile:
Basically the Hermitian subalgebra of the whole C*-algebra is meant to correspond to measuring devices.
 
  • #35
DarMM said:
Basically the Hermitian subalgebra of the whole C*-algebra is meant to correspond to measuring devices.
This goes back to my earlier gripe -- there's nothing physically stopping me from making a measuring device that outputs complex numbers. Singling out those elements whose anti-Hermetian part is zero as being more "real" is just an extension of the old bias that the complex numbers with zero imaginary part are somehow more real than the rest of them.
 
  • #36
Hurkyl said:
This goes back to my earlier gripe -- there's nothing physically stopping me from making a measuring device that outputs complex numbers. Singling out those elements whose anti-Hermetian part is zero as being more "real" is just an extension of the old bias that the complex numbers with zero imaginary part are somehow more real than the rest of them.
I wouldn't so much see the "reality" of the measured output as being the reason for requiring Hermiticity. Of course if I take an observable A and an observable B, I can measure A + iB by just getting their values and put them into a complex number. Rather it has more to do with Unitarity. If an observable is not Hermitian then the transformation associated with it is not Unitary and it does not represent a good quantum number or even allow sensible quantum evolution. For example if H, the Hamiltonian wasn't Hermitian then time evolution wouldn't be Unitary, which would make the theory collapse. Similarly for momentum, linear and angular, rotations and translations wouldn't be unitary.
Hence Hermitian operators represent our observables, because only they represent good quantum numbers. For example only then will we be sure that when we obtain A = a that we are in a specific state by the spectral theorem.
Another example would be that measuring A + iB only really makes sense if A and B are compatible observables. So if the Hamiltonian, Linear Momentum and Angular Momentum have to be Hermitian, functions of them essentially exhaust all operators.
There are other reasons for Hermiticity, which I can go into if you want.
 
  • #37
It's clear why translation and other automorphisms should be unitary: they have to preserve the C* structure. And it's clear why the inner automorphisms -- those of the form T \mapsto U^* T U -- require U to be unitary element.


But (a priori, anyways) that has absolutely nothing to do with whether an element of the C*-algebra should correspond to a measuring device.



Actually, you bring up an interesting example. If you have a one-parameter family of unitary transformations U(t) -- such as time translation -- the corresponding infinitessimal element U'(0) is anti-Hermitian, not Hermitian. That we divide out by i to get a Hermitian element appears to me to be for no reason deeper than "people like Hermitian elements".
 
  • #38
Hi.

Fredrik said:
Eigenvalue 0 doesn't cause any additional complications at all, so it doesn't need to be handled separately.
Thanks a lot, Fredrik. Now I am fine with this eigenvalue 0 concern.

Now going back to my original question
sweet springs said:
Hi,
In 9.2 of my old textbook Mathematical Methods for Physicists, George Arfken states,
------------------------------------------------------
1. The eigenvalues of an Hermite operator are real.
2. The eigen functins of an Hermite operator are orthogonal.
3. The eigen functins of an Hermite operator form a complete set.*
* This third property is not universal. It does hold for our linear, second order differential operators in Strum-Liouville (self adjoint) form.
------------------------------------------------------
Advice on * of 3. , showing some "not forming a complete set" examples, are open and appreciated.

Regards.
 
  • #39
Hurkyl said:
It's clear why translation and other automorphisms should be unitary: they have to preserve the C* structure. And it's clear why the inner automorphisms -- those of the form T \mapsto U^* T U -- require U to be unitary element.


But (a priori, anyways) that has absolutely nothing to do with whether an element of the C*-algebra should correspond to a measuring device.



Actually, you bring up an interesting example. If you have a one-parameter family of unitary transformations U(t) -- such as time translation -- the corresponding infinitessimal element U'(0) is anti-Hermitian, not Hermitian. That we divide out by i to get a Hermitian element appears to me to be for no reason deeper than "people like Hermitian elements".
Well let's concentrate on just the Hamiltonian. As you said we could work with A = iH, but we choose to work with A = iH. However this isn't really a very interesting case, it's similar to the case I described before in terms of sticking an i in front. So for example A + iB is fine as an observable, get a machine that measures both and add them together inside the machine and the machine will have measured a + ib. Nothing wrong with that. However these are trivial complex observables, formed from Hermitian observables anyway. What about a genuine complex observable like a non-Hermitian Hamiltonian which can produce eigenstates like E + i\Gamma? The problem is that such eigenvalues are actually describe decaying non-observable particles and there won't be a conservation of probability.
My basic idea is that while we can have things like iH, it is difficult to justify an arbitrary non-Hermitian operator. In the case of the Hamiltonian its because of the loss of conservation of probability, in the case of other operators it's usually because non-Hermitian observables don't form an orthogonal basis and hence aren't good quantum numbers.

However I can see that what I've said is basically an argument as to why non-Hermitian operators would be bad things to measure. What I haven't explained is why they actually can't be measured physically. I'll explain that in my next post since it takes a bit of work to set up.
 
  • #40
However these are trivial complex observables, formed from Hermitian observables anyway.
For the record, by this definition of "trivial", all elements of a C*-algebra are trivial: you can compute the real and imaginary parts just like an ordinary scalar:
X = (Z + Z*) / 2
Y = (Z - Z*) / (2i)​
giving
X* = X
Y* = Y
Z = X + iY​
 
  • #41
Hurkyl said:
For the record, by this definition of "trivial", all elements of a C*-algebra are trivial: you can compute the real and imaginary parts just like an ordinary scalar:
X = (Z + Z*) / 2
Y = (Z - Z*) / (2i)​
giving
X* = X
Y* = Y
Z = X + iY​
Yeah, true. Bad example on my part, hopefully I can explain why we restrict ourselves to Hermitian observables in the next post. Once I have outlined the idea from the algebraic point of view hopefully we can have a more fruitful discussion.
Also I should say that in the relativistic context not all Hermitian operators are observables.
 
  • #42
Okay, when we make a measurement, a quantum mechanical object in the state \psi interacts with a classical measuring apparatus to record a value of some quantity \mathbb{A}. Mathematically this quantity is represented by an operator A. All the statistics for the observable such as the expectation, standard deviation, uncertainty e.t.c. can be worked out from the state and the observable. Let's take the expectation value, in your opinion should the expectation be represented as
\langle \psi, A\psi\rangle
or
\langle A\psi, \psi\rangle
Which one of these should represent an experiment to measure A?
 
  • #43
DarMM said:
Let's take the expectation value, in your opinion should the expectation be represented as
\langle \psi, A\psi\rangle
or
\langle A\psi, \psi\rangle
Which one of these should represent an experiment to measure A?
Well, it would depend on how we chose to use the Hilbert space to represent states.

I'm going to go with the former, though. If \psi is a ket corresponding to the expectation functional \rho, then I prefer to have \rho(A) = \psi^* A \psi, which corresponds to the convention relating duals to inner products I assume we're using.
 
  • #44
Hurkyl said:
Well, it would depend on how we chose to use the Hilbert space to represent states.

I'm going to go with the former, though. If \psi is a ket corresponding to the expectation functional \rho, then I prefer to have \rho(A) = \psi^* A \psi, which corresponds to the convention relating duals to inner products I assume we're using.
Funnily enough I should say, before I go on, that some people use my example above to argue why observables should not be Hermitian. That is they feel that physics should give the same answers regardless of which choice you use \langle \psi, A\psi\rangle or \langle A\psi, \psi\rangle. Or to put it in loose language "experiments cannot test the inner product".

Anyway on to the more important fact. It is an observed consequence of atomic measurement that if we measure a physical quantity and then measure that quantity again with no other quantities measured in between, then the chance of us obtaining the same answer is 100%. Given that this is fact of measurement how can we model it? Well if we obtained the value a we are in the state |a\rangle. Then since we know that we have no chance of measuring another value b for the same observable we want the probability to vanish for transition from |a\rangle to |b\rangle. That is we want \langle b,a\rangle = 0. So in order to match experiment observables must be represented by operators whose eigenvectors are orthogonal. Would you agree?
(Please tell me if something is incorrect.)
 
  • #45
I ground through some calculations, and I'm pretty sure that transition amplitude from \psi to b ought to be
the coefficient of |b\rangle in the representation of |\psi \rangle relative to the eigenbasis​
and not
the inner product of b with \psi.​

Of course, if you have an orthonormal eigenbasis, they are the same.
 
  • #46
Hurkyl said:
I ground through some calculations, and I'm pretty sure that transition amplitude from \psi to b ought to be
the coefficient of |b\rangle in the representation of |\psi \rangle relative to the eigenbasis​
and not
the inner product of b with \psi.​

Of course, if you have an orthonormal eigenbasis, they are the same.
Really, why do you say? Perhaps I'm missing something, but I thought the usual definition of the transition probability was \langle b,a\rangle. How did you calculate what the transition probability was? Maybe I'm just being silly though!
 
  • #47
Well, the heuristic calculation I went through was as follows:

First, I want to make a toy example of a unitary operator that collapses the state in question. I chose the following one for no particular reason other than it was simple:
T|a,e_b> = |a,e_{b+a}>​
The Hilbert state space here is the tensor product of the state space we are interested in, with a basis labeled by the eigenvalues of whatever operator we're interested in, and another state space representing a toy environment, with basis states labeled by complex numbers. ("e" for "environment")

I chose a generic pure state in ket form:
|\psi\rangle = \sum_a c(a) |a\rangle​
computed the density matrix of the state:
T(|\psi \rangle \otimes |0 \rangle)​
and took the partial trace to get the resulting density matrix:
\sum_{a,b} c(b)^* c(a) \langle e_a | e_b \rangle\, |a\rangle\langle b|​
Since this evolution was supposed to collapse into a mixture of the eigenstates, I convinced myself that implies the environment states do need to be orthogonal, giving the density matrix:
\sum_{a} |c(a)|^2 |a\rangle\langle a|​
which is the statistical mixture that has probability |c(a)|^2 of appearing in state |a\rangle.

This toy seems reasonable since it gives the statistical mixture I was expecting, and eigenstates (e.g. |a\rangle) remain fixed, so the mixture generally remains stable.

So, if transition probabilities make sense at all, the transition probability from |\psi\rangle to |a\rangle has to be |c(a)|^2 -- in other words, the right computation for transition amplitude is the "coefficient of |a\rangle" function, rather than the "inner product with |a\rangle" function.
 
  • #48
George Jones said:
strangerep said:
A "linear combination" in an arbitrary vector space can certainly be an
infinite sum.
For an arbitrary vector space, what does "infinite sum" mean? There is only one topology, the Euclidean topology, that can be given to a finite-dimensional vector space, but an infinite-dimensional vector space can be given various topologies.

That's why I tried to distinguish such an arbitrary vector space
from a Hilbert space in my post.

I should possibly have said formal linear combination. I was thinking of the
"universal" space mentioned in Ballentine section 1.4.

Certainly, one can't do very much useful stuff in such an arbitrary vector space
before equipping it with a topology.
 
  • #49
Fredrik said:
[...] I think it's more appropriate to define a linear combination to only
have a finite number of terms.
Consider the usual kind of inf-dim Hilbert space which has an
orthonormal basis consisting of an infinite number of vectors.
In general, an arbitrary vector in that space can be expressed
as an infinite sum over the basis vectors. Surely such a sum
qualifies as a linear combination?

This is why: If V is a vector space over a field F, and S is a subset of V, the "subspace generated by S" (or "spanned" by S) can be defined as any of the following

a) the smallest subspace that contains S
b) the intersection of all subspaces that contain S
c) \Big\{\sum_{i=1}^n a_i s_i|a_i\in\mathbb F, s_i\in S, n\in\mathbb N\Big\}

These definitions are all equivalent, and the fact that every member of the set defined in c) can be expressed as \sum_{i=1}^n a_i s_i, with n finite, seems like a good reason to define a "linear combination" as having only finitely many terms. [...]

One can also find inf-dim subspaces in general, in which case those arguments about
finite sums don't apply.
 
  • #50
strangerep said:
...an infinite sum over the basis vectors. Surely such a sum
qualifies as a linear combination?
The theorem I mentioned is valid for arbitrary vector spaces. (I'm including the proof below). It implies that if we define "linear combination" your way, the following statement is false:

The subspace generated (=spanned) by S is equal to the set of linear combinations of members of S.

So let's think about what you said in the text I quoted. The definition of "subspace generated by" and the theorem imply that a vector expressed as

x=\sum_{n=1}^\infty \langle e_n,x\rangle e_n

with infinitely many non-zero terms does not belong to the subspace generated by the orthonormal basis. :confused: That's odd. I didn't expect that.

I think the explanation is that terms like "linear combination" and "subspace generated by" were invented to be useful when we're dealing with arbitrary vector spaces, where infinite sums may not even be defined. And then we stick to the same terminology when we're dealing with Hilbert spaces.

I haven't tried to prove it, but I'm guessing that the subspace spanned by an orthonormal basis is dense in the Hilbert space, and also that it isn't complete. But vectors like the x mentioned above can be reached (I assume) as a limit of a sequence of members of the subspace generated by the basis. (A convergent sum of the kind that appears on the right above is of course a special case of that).

strangerep said:
One can also find inf-dim subspaces in general, in which case those arguments about
finite sums don't apply.
The theorem holds for infinite-dimensional vector spaces too. The proof is very easy. Let V be an arbitrary vector space, and let S be an arbitrary subset. Define \bigvee S to be the intersection of all subspaces W_\alpha such that S\subset W_\alpha. I'll write this intersection as

\bigvee S=\bigcap_\alpha W_\alpha

Define W to be the set of all linear combinations of members of S. (Both here and below, when I say "linear combination", I mean something with a finite number of terms).

W=\Big\{\sum_{i=1}^n a_i s_i|a_i\in\mathbb F, s_i\in S, n\in\mathbb N\Big\}

I want to show that W=\bigvee S. First we prove that W\subset\bigvee S.

Let x be an arbitrary member of W. x is a linear combination of members of S, but S is a subset of every W_\alpha. So x is a linear combination of members of W_\alpha for every \alpha. The W_\alpha are subspaces, so that implies that x is a member of every W_\alpha. Therefore x\in\bigvee S[/tex], and that implies W\subset\bigvee S.<br /> <br /> Then we prove that \bigvee S\subset W. It&#039;s obvious from the definition of W that it&#039;s closed under linear combinations. That means that it&#039;s a subspace. So it&#039;s one of the terms on the right in<br /> <br /> \bigvee S=\bigcap_\alpha W_\alpha<br /> <br /> That implies that \bigvee S\subset W.
 
  • #51
Hurkyl said:
Well, the heuristic calculation I went through was as follows:...
Now, I will add there's I don't like something about my derivation, but I haven't managed to place my finger on it.
 
  • #52
Hurkyl said:
Now, I will add there's I don't like something about my derivation, but I haven't managed to place my finger on it.
I found it -- my T wasn't unitary. I'll have to work up a better toy example.
 
  • #53
Fredrik said:
[...]
So let's think about what you said in the text I quoted. The definition of "subspace generated by" and the theorem imply that a vector expressed as

x=\sum_{n=1}^\infty \langle e_n,x\rangle e_n

with infinitely many non-zero terms does not belong to the subspace generated by the orthonormal basis. :confused: That's odd. I didn't expect that.

[...]

I get the feeling we've been talking at crossed purposes. When I read something
like "X is a subspace of V", I've been tacitly assuming X the same kind of space
as whatever V is. I.e., if V is a vector space, then X is also a vector space, or if
V is a Hilbert space, then X is also a Hilbert space, etc. But I probably shouldn't
be assuming that's what you meant.

Such ambiguity is probably the source of any misunderstandings.
 
  • #54
You're right, I didn't think about the fact that a "subspace" of a Hilbert space should be complete. I meant subspace in the vector space sense. I'm a bit busy right now, so I haven't had time to think about how or if that changes the stuff I said.
 
  • #55
Hi.
Let A and B be each Hermitian and OBSERVABLE in the sense that whose eigensubspaces contain all the maximal orthogonal sets, i.e. basis.
A+B is Hermite. Is A+B OBSERVABLE? e.g. X + h' P^-1.
Can any Hermitian be diagonalized?
Regards.
 
Last edited:
  • #56
I assume that you're trying to say that there's a basis for the Hilbert space that only contains eigenvectors of the operator. (If you're going to talk about eigenspaces, you'll have to say that the direct sum of the eigenspaces is the entire Hilbert space).

If the above is true for A and B separately, is it also true for A+B? Yes, it is, because A+B is hermitian too.


(And for the record, I still think that's a bad definition of "observable").
 
  • #57
Hi, Fredrik. Thank you so much.
With help of your teachings I could confirm that Hamiltonian of a particle in any artificial potential, say H=P^2/2m + V(X), has eigenstates and is OBSERVABLE whatever V(X) is.
Regards.
 
  • #58
Yes. H must be hermitian because exp(-iHt) must be unitary.
 
  • #59
Hi.
I still have a little concern on the description below. But I cannot imagine an operator that is Hermitian but not OBSERVABLE i.e. the direct sum of whose eigensubspaces is the whole space. Now I stop wondering about this subject. Thanks a lot to you all. :smile:
In 9.2, Mathematical Methods for Physicists, George Arfken
------------------------------------------------------
1. The eigenvalues of an Hermite operator are real.
2. The eigen functins of an Hermite operator are orthogonal.
3. The eigen functins of an Hermite operator form a complete set.*
* This third property is not universal. It does hold for our linear, second order differential operators in Strum-Liouville (self adjoint) form.
------------------------------------------------------
 
Last edited:

Similar threads

  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 14 ·
Replies
14
Views
2K
  • · Replies 17 ·
Replies
17
Views
1K
  • · Replies 15 ·
Replies
15
Views
547
  • · Replies 3 ·
Replies
3
Views
1K
  • · Replies 24 ·
Replies
24
Views
2K
Replies
6
Views
1K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 5 ·
Replies
5
Views
2K
Replies
9
Views
2K