Exploring the Fundamental Postulates of QM: Are They Truly Ad-Hoc and Strange?

  • #51
stevendaryl said:
In my opinion, calling an observable of a C*-algebra an "equivalence class of measuring devices" is more suggestive than rigorous. I would think that if one really wanted to seriously talk about equivalence classes, then one would have to

  1. Define what a "measuring device" is.
  2. Define an equivalence relation on measuring devices.
  3. Define the operations on measuring devices (addition, multiplication, scaling, or whatever).
  4. Prove that the equivalence relation is a congruence with respect to those operations.

I don't think you can really do that in a noncircular way, because to make sense of the claim that a particular device is a measuring device for the z-component of spin angular momentum of some particle, you would need to assume some kind of dynamics whereby the device interacts with the particle so that its state evolves to a persistent record of the z-component of the spin angular momentum. You need to have a theory of interactions before you can ever know that something is a measuring device. So it's a bit weird to put in equivalence classes of measuring devices at the beginning, as opposed to having them come out of the theory.


A full definition of a theory must include statements that tell us how to interpret the mathematics as predictions about measurement results. These statements, called "correspondence rules", must tell us what sort of devices we're supposed to use. This is where things get complicated.

Let's say that we want to write down the correspondence rules for (say) the theory of classical point particles in Minkowski spacetime. One of the rules must specify what a clock is. This is a problem. We can't just say that a clock is a device that measures time, because "time" is defined by the theory we're trying to define. The solution is to define a clock by explicit instructions on how to build one.

In principle those instructions can be written so that they can be followed by people who don't know any physics at all, but I can't even imagine what they would look like if we write them that way.

This is still pretty weird, because the best clocks are designed using SR, QM and a lot more. I'm not sure we absolutely need to address that issue, but I see one way that it can be addressed: We define a hierarchy of theories. In the level-0 theories, we use very simple descriptions of measuring devices. Then for each positive integer n, when we define the level-n theories, we make sure that the instructions in the correspondence rules can be understood by people who understand level-(n-1) theories and have access to level-(n-1) measuring devices.

As you can see this is all really complicated, and this is just a discussion of what it takes to completely write down the definition of a good theory (something that certainly has never been done). But I think it's clear that we can at least avoid circularity in the definition of the theory.

I have to go, so I don't have time to address the issue of circularity in the algebraic approach. Maybe later. (I don't think there is any circularity there).
 
Physics news on Phys.org
  • #52
When Schrodinger posited his wavefunction, it came at a time when physical theory was transitioning from physical Newtonian space into "metaphysical" Hilbert space. Heisenberg was only interested in developing a matrix algebra of observed quantities. One of the uses of the wavefunction is to use its solution in three dimensions as the "orbitals" of electrons. This is the exact same concept as the Bohr model, except that the "orbitals" of the wavefunction are quite a bit more convoluted.

Both the Bohr and Schrodinger models can be viewed as classical in that they are are both (at least in theory) represented as existing in classical space. This is what it means for something to be a "model". Heisenberg was always adamant that his ideas never have any connection to the classical spacetime models of pre-20th century physical theory.

I think what makes things so strange is simply that connections were eventually made between the Heisenberg and the Schrodinger "ontologies". As Charles mentioned above, Bohr was the great philosopher who made it his mission to get everyone to put aside their personal prides for the sake of the larger goal of getting a cohesive vision on the table. Born was able to get rid of the classical model by squaring the wavefunction. And Dirac finally gave everything a formal language with his new bra-ket notation.

This is all quite a lot for mere mortals to put into proper perspective...
 
  • #53
stevendaryl said:
In my opinion, calling an observable of a C*-algebra an "equivalence class of measuring devices" is more suggestive than rigorous.
More suggestive than rigorous...yes, I suppose so. If we want to do it rigorously, we must start by stating a definition of "theory" that's general enough to include all the classical and all the quantum theories. We can then define terms like "state" and "observable" in a way that's both rigorous and theory-independent (in the sense that the same definition applies to all the classical theories, all the quantum theories, and more).

I spent some time thinking about how to do these things a couple of years ago. I didn't keep at it long enough to work everything out, but I feel very strongly that it can be done. The first step is to provide some motivation for a general definition of "theory". This is of course impossible to do rigorously, but the main ideas are very simple and natural. (Actually, what we want to define here isn't a theory of physics in the sense of my previous posts in this thread. It's just the purely mathematical part of such a theory, not including any correspondence rules. So maybe we should use some other term for it, but "theory" will have to do in this thread).

The idea that I used as the starting point is that a theory must be able to assign probabilities to statements of the form
"If you use the measuring device \delta on the object \pi, the result will be in the set E".​
These statements can be identified by the triples ##(\delta,\pi,E)##. This means that associated with each theory, there are sets ##\Delta,\Pi,\Sigma## and a function
$$P:\Pi\times\Delta\times\Sigma\rightarrow[0,1],$$ such that the maps ##E\mapsto P(\delta,\pi,E)## are probability measures. Note that this implies that ##\Sigma## is a σ-algebra. I call elements of the set ##\Delta## "measuring devices" and elements of the set ##\Pi## "preparations".

After these simple observations and conjectures, we are already very close to being able to write down a definition that we can use as the starting point for rigorous proofs. There are some subtleties that we have to figure out how to deal with before we write down a definition, like what happens if the measured object ##\pi## is too big to fit in the measuring device ##\delta##? I'm not going to try to work out all such issues here, I'm just saying that they look like minor obstacles that are unlikely to prevent us from finding a satisfactory definition.

Now let's jump ahead a bit and suppose that we have already written down a satisfactory definition, and that the sets and functions I've mentioned are a part of it. Then we can use the function P to define equivalence classes and terms like "state" and "observable". This function implicitly defines several others, like the maps E\mapsto P(\pi,\delta,E) already mentioned above. We will be interested in the functions that are suggested by the following notations:
\begin{align}
P(\pi,\delta,E)=P_\pi(\delta,E)=P^\delta(\pi,E)=P_\pi^\delta(E)
\end{align} We use the P_\pi and P^\delta functions to define equivalence relations on \Pi and \Delta: \begin{align*}<br /> &amp;\forall \pi,\rho\in\Pi\qquad &amp;\pi \sim \rho\quad &amp;\text{if}\quad P_\pi=P_\rho\\<br /> &amp;\forall \delta,\epsilon\in\Delta\qquad &amp;\delta \sim \epsilon\quad &amp;\text{if}\quad P^\delta=P^\epsilon<br /> \end{align*}
The sets of equivalence classes are denoted by \mathcal S and \mathcal O respectively. The members of \mathcal S=\Pi/\sim are called states, and the members of \mathcal O =\Delta/\sim are called observables. The idea behind these definitions is that if two members of the same set can't be distinguished by experiments, the theory shouldn't distinguish between them either.

stevendaryl said:
I would think that if one really wanted to seriously talk about equivalence classes, then one would have to

  1. Define what a "measuring device" is.
As you can see, I don't agree with this point. We only have to define the term "theory" in such a way that every theory is associated with a set whose elements we can call "measuring devices".

If we continue along the lines I've started above, we will not automatically end up with C*-algebras. What I'm describing can (probably) be thought of as a common starting point for both the algebraic approach and the quantum logic approach. So we can proceed in more than one way from here. We can define operations on the set of observables that turn it into a normed vector space, and then think "wouldn't it be awesome if this is a C*-algebra?", or we can keep messing around with equivalence classes and stuff until we find a lattice, and then think "wouldn't it be awesome if this is orthocomplemented, orthomodular, and whatever else we need it to be?".

I seems to me that the reason why we don't get the most convenient possibility to appear automatically, is that we started with a definition that's "too" general. It doesn't just include all the classical theories and all the quantum theories, it includes a lot more. So if we want to consider only classical and quantum theories, we need to impose additional conditions on the structure (a normed vector space or a lattice), that gets rid of the unwanted theories.

stevendaryl said:
I don't think you can really do that in a noncircular way,
I think the approach I have described doesn't have any circularity problems.

stevendaryl said:
But the C*-algebra approach sure seems to single out measurements (or observables) as being something different. As I said, the fact that some interaction is a measurement of some quantity is not what you start with, it's a conclusion. There's a long chain of deductions involved in reaching that conclusion.
The way I see it, the chain of deductions that lead to this conclusion is based only on the concept of "falsifiability". And the conclusion provides the motivation for a definition of the term "theory of physics".
 
Last edited:
  • #54
Fredrik, this is possibly one of the most interesting posts I've read on this forum. It's too bad you never completely worked everything out. It would love to read something like that.
 
  • #55
micromass said:
Fredrik, this is possibly one of the most interesting posts I've read on this forum. It's too bad you never completely worked everything out. It would love to read something like that.
Thanks. I'm glad you liked it. Maybe I'll have another go at completing it soon.
 
  • #56
stevendaryl said:
People informally say that A B means "first measure B", then measure A", [...]
As Ballentine points out somewhere in his textbook, such an interpretation of a product of operators is also clearly wrong.

Consider the case ##A = \sigma_x##, ##B=\sigma_y## (where the ##\sigma##'s are the usual Pauli matrices). Then we have ##AB = i\sigma_z##, but "a measurement of spin along the x-axis followed by a measurement of spin along the y axis" is in no sense relatable to a single measurement of spin along the z axis.

Moreover, even if ##A,B## are both hermitian, we could have ##(AB)^* = B^* A^* = BA \ne AB## in general. Hence ##AB## does not necessarily qualify as an observable in the ordinary sense of an hermitian operator.

Products of operators are better understood in the context of the full dynamical group of the system under study. E.g., in terms of the universal enveloping algebra (or maybe Poisson algebra) associated with the generators of that group. One constructs unitary representations of the group.
 
Last edited:
  • #57
HomogenousCow said:
I agree but it baffles me why this model works, none of the postulates are directly motivated by experimental evidence, only after some deep digging do we find that they agree with interference and other observations.

I would like to suggest you get a hold of Ballentine - Quantum Mechanics - A Modern Development. There you will find the correct basis of QM - it really rests on two axioms. Stuff like Schrodinger's equation etc follows from the the POR exactly the same as SR does. The second axiom he uses, which is basically Born's Rule follows from the first axiom if you accept non contextuality (which is highly intuitive mathematically) via Gleason's theorem so one can argue it is really based on one axiom with a bit of other stuff added.

The issue is can the two axioms be presented in an intuitive way? I believe it can - check out:
http://arxiv.org/pdf/0911.0695v1.pdf

It would seem some fairly general and intuitive considerations leads either to bog standard probability theory or QM - with QM being singled out if you want continuous transformations between pure states or entanglement - either one is enough to uniquely determine QM as the correct model.

Thanks
Bill
 
Last edited:
  • #58
dextercioby said:
Everything in life, including science, is a matter of subjectivity. You like that, you don't like that, what's not intuitive to you is for someone else. The more subjective you are, the less you know about a certain topic. GR and Quantum Mechanics are both equally valid generalizations of classical mechanics. Formulate classical mechanics in such a way that both GR and QM are the natural extensions of it. If you think that CM is F=ma + forces add together like vectors and the force of 1 on 2 is the opposite of the force of 2 on 1, then you're not ready for GR, not ready for QM and seeing the (diluted or not) axiomatization of QM would make you say: <This is weird. Where did it come from?>.

Exactly. What is reasonable to one person is crazy to another. I have found approaches to QM that for me make it seem quite reasonable.

dextercioby said:
Advice: read more and don't be afraid of mathematics. The more math you know, the more logical will the advanced physics look to you.

Yea - that seems to be a big problem for some. Those 'reasonable' approaches often use advanced math which can be a turn off - which of course it shouldn't be - physics is not math but is written in the language of math so its hardly surprising it takes its most elegant and transparent form within that framework.

Thanks
Bill
 
  • #59
bhobba said:
Yea - that seems to be a big problem for some. Those 'reasonable' approaches often use advanced math which can be a turn off - which of course it shouldn't be - physics is not math but is written in the language of math so its hardly surprising it takes its most elegant and transparent form within that framework.

As I said already, I don't think that the strangeness of quantum mechanics has anything to do with the difficulty of the math. To give some counter-examples, I think that General Relativity or statistical mechanics can be just as difficult, mathematically. I really think that it is the singling out "observables" as a fundamental, irreducible aspect of the world that is strange.
 
  • #60
bhobba said:
The issue is can the two axioms be presented in an intuitive way? I believe it can - check out:
http://arxiv.org/pdf/0911.0695v1.pdf

Thanks for that reference. To me, the strangeness is already put in at the very beginning, when a "measurement" is given fundamental status in the axioms. As I said, a "measurement" is not a fundamental, atomic entity, but is a special kind of interaction whereby the state of one system (the observer, or measuring device) becomes correlated, in a persistent way, with the state of another system (the thing being observed or measured). The discussion, where one talks about "reliably distinguishing" states, is already, it seems to me, making a division between the world and the thing that is studying the world. Of course, that distinction is certainly there when you have a scientist doing experiments, but I always felt that that was a matter of how we interpreted what was going on--at the level of the laws of physics, there's no fundamental distinction between scientist and experiment.
 
  • #61
Fredrik said:
The idea that I used as the starting point is that a theory must be able to assign probabilities to statements of the form
"If you use the measuring device \delta on the object \pi, the result will be in the set E".​

So your approach is to let "measuring device" be an abstract term. But what is supposed to be the interpretation? Suppose we have as a simple case a universe consisting of nothing but a single spin-1/2 particle fixed in place (so the only degrees of freedom are from spin). I assume that the "observables" in this case are associated with the set of 2x2 hermitian matrices. Which means, in terms of Pauli spin matrices \sigma_i, that they are of the form:
A + B_i \sigma_i, where A, B_x, B_y, B_z are 4 real numbers. So for this toy theory, each such matrix is a measuring device?
 
Last edited:
  • #62
stevendaryl said:
So your approach is to let "measuring device" be an abstract term. But what is supposed to be the interpretation? Suppose we have as a simple case a universe consisting of nothing but a single spin-1/2 particle fixed in place (so the only degrees of freedom are from spin). I assume that the "observables" in this case are associated with the set of 2x2 hermitian matrices. Which means, in terms of Pauli spin matrices \sigma_i, that they are of the form:
A + B_i \sigma_i, where A, B_x, B_y, B_z are 4 real numbers. So for this toy theory, each such matrix is a measuring device?
The set Δ whose members I call "measuring devices" contains elements that correspond to the actual measuring devices mentioned by the theory's correspondence rules. But I do not require that every element of Δ corresponds to an actual measuring device. We can take Δ to be a larger set, if that's convenient.

In a quantum theory defined by a 2-dimensional Hilbert space, the set of self-adjoint operators is our Δ/~ (i.e. the set of all equivalence classes of measuring devices). This is a 4-dimensional vector space over ℝ, that's spanned by ##\{\sigma_1,\sigma_2,\sigma_3,I\}##. The sigmas correspond to measuring devices that measure spin in one of three orthogonal directions. The identity matrix corresponds to a measuring device that always gives us the result 1, no matter how the system was prepared before the measurement. Since every self-adjoint operator is a linear combination of these four, every self-adjoint operator corresponds to an actual measuring device (assuming that linear combinations of observables make sense).

Regarding the meaning of linear combinations... I defined scalar multiplication earlier. I haven't really thought addition through. Strocchi appears to be doing something like this: If we denote the expectation value of an observable X by E(X|s) when the system is in state s, then addition can be defined by saying that A+B is the observable such that E(A+B|s)=E(A|s)+E(B|s) for all states s. (I haven't verified that this definition works).
 
  • #63
stevendaryl said:
To me, the strangeness is already put in at the very beginning, when a "measurement" is given fundamental status in the axioms. As I said, a "measurement" is not a fundamental, atomic entity, but is a special kind of interaction whereby the state of one system (the observer, or measuring device) becomes correlated, in a persistent way, with the state of another system (the thing being observed or measured).
As you know, physics obtains its knowledge by measurements. So first of all, the resulting theories are theories about measurements. In QM, it is not straightforward how to extrapolate the theory about measurements to a theory about what "really happens". From the viewpoint of common forms of realism, this is a problem of course. But why should we expect such a straightforward extrapolation in the first place?

Also conceptually, the Many Worlds interpretation is quite straightforward and tells us what really happens. However, it's still hard to accept from the viewpoint of naive realism.
 
  • #64
kith said:
As you know, physics obtains its knowledge by measurements.

Sure.

So first of all, the resulting theories are theories about measurements.

I don't think that follows at all. That's like saying: "Nowadays, many people learn about physics over the internet. So for them, a theory of physics is a theory of web browsers."

We learn about physics by measurements, but measurements are not the subject of physics. (Well, there can certainly be a subfield of physics, the theory of measurement, but that's not all of physics.) We use measurements to figure out how the world works, and then we apply that knowledge in situations where there are no measurements around--such as the Earth prior to the formation of life, or inside a star, or whatever.

I absolutely reject the assumption that a theory of physics is a theory of measurement.
 
  • #65
stevendaryl said:
To me, the strangeness is already put in at the very beginning, when a "measurement" is given fundamental status in the axioms. As I said, a "measurement" is not a fundamental, atomic entity, but is a special kind of interaction whereby the state of one system (the observer, or measuring device) becomes correlated, in a persistent way, with the state of another system (the thing being observed or measured).

Then, I suspect, to you, that entanglement basically leads to QM would be a very pertinent aspect.

My view is a few approaches with reasonable foundations lead to QM but unfortunately, like the paper I linked, require a certain amount of mathematical sophistication such as the Schur-Auerbach lemma from group theory. Unfortunately there are some people who don't like this mathematical aspect of physical theories and in some quarters there is a resistance to it with for example claims SR is simply math and can't represent physical reality. I had long discussions (if that's what you would call them) with people of that bent when I posted a lot on sci.physics.relativity - they just simply can't get the idea that physics is not about easily visualizeable pictures they carry around in their head.

Thanks
Bill
 
  • #66
stevendaryl said:
That's like saying: "Nowadays, many people learn about physics over the internet. So for them, a theory of physics is a theory of web browsers."
This analogy is valid if you make the assumption that measurements uncover an independent reality. But this is exactly the assumption I questioned in my previous post.
 
  • #67
kith said:
As you know, physics obtains its knowledge by measurements. So first of all, the resulting theories are theories about measurements.

Yes of course. But that in itself raises a fundamental issue - measurement apparatus are classical objects and QM is a fundamental theory about the constituents of those classical objects so we have a 'cut' in how we view nature right at the foundations of QM. This leads to stuff like the Von Newmann regress and the introduction of consciousness causes collapse most would think a bit too far out to be taken seriously. My view is a fully quantum theory of measurement is required and indeed much progress in that area has been made but a few issues still remain such as proving the basis singled out by decoherence does not depend on how the system is decomposed. I believe we are not far away from a full resolution but until all the i's are dotted at t's crossed I for one still think some mystery remains. And who knows - dotting the i's and crossing the t's may show up something truly surprising.

Thanks
Bill
 
  • #68
kith said:
This analogy is valid if you make the assumption that measurements uncover an independent reality. But this is exactly the assumption I questioned in my previous post.

Bingo - we have a winner. That is the rock bottom foundational issue with QM IMHO.

Thanks
Bill
 
  • #69
Bill, similar to your thoughts, I think that the universal wavefunction and decoherence give a quite complete realistic picture. It is just that I also see the appeal in the C* approach which is much closer to the scientific practise than unobservable entities like the universal wavefunction.
 
Last edited:
  • #70
bhobba said:
Yes of course. But that in itself raises a fundamental issue - measurement apparatus are classical objects and QM is a fundamental theory about the constituents of those classical objects so we have a 'cut' in how we view nature right at the foundations of QM. This leads to stuff like the Von Newmann regress and the introduction of consciousness causes collapse most would think a bit too far out to be taken seriously. My view is a fully quantum theory of measurement is required and indeed much progress in that area has been made but a few issues still remain such as proving the basis singled out by decoherence does not depend on how the system is decomposed. I believe we are not far away from a full resolution but until all the i's are dotted at t's crossed I for one still think some mystery remains. And who knows - dotting the i's and crossing the t's may show up something truly surprising.
What do you mean by a fully quantum theory of measurement, if QM isn't one already? (Keep in mind that QM includes decoherence). And what is it required for?

The von Neumann regress, if you mean what I think you mean, doesn't have anything to do with the consciousness causes collapse idea. The former is just the observation about what a theory is, and the latter is at best a wild speculation about reality.

I think the idea that the basis is independent of the decomposition is as likely to be true as the idea that 2x is independent of x.
 
  • #71
Fredrik said:
What do you mean by a fully quantum theory of measurement, if QM isn't one already? (Keep in mind that QM includes decoherence). And what is it required for?

A theory of measurement that does not include the a priori existence of classical measurement devices like Copenhagen does. It is only of recent times such theories have emerged - but as yet have not been full worked out eg my understanding is that the emergence of a classical domain from QM is not quite 100% complete - we are almost there - but not quite - at least that's what I have read.

Fredrik said:
The von Neumann regress, if you mean what I think you mean, doesn't have anything to do with the consciousness causes collapse idea. The former is just the observation about what a theory is, and the latter is at best a wild speculation about reality.

Von Neumann was one of the first, or maybe even the first, to examine the measurement process fully quantum mechanically. What that showed is where the wavefunction collapse occurs could be placed anywhere and if we keep following it all the way back to the observer it is only at consciousness something different comes into it. To some such as Wigner this is where they placed the collapse:
http://en.wikipedia.org/wiki/Interp...rpretation:_consciousness_causes_the_collapse

'In his treatise The Mathematical Foundations of Quantum Mechanics, John von Neumann deeply analyzed the so-called measurement problem. He concluded that the entire physical universe could be made subject to the Schrödinger equation (the universal wave function). He also described how measurement could cause a collapse of the wave function. This point of view was prominently expanded on by Eugene Wigner, who argued that human experimenter consciousness (or maybe even dog consciousness) was critical for the collapse, but he later abandoned this interpretation'

I have an aged copy of Von Newmann's text, and while it been years since I have read it (its actually one of the first books I learned QM from because it was mathematically more in line with the Hilbert spaces I studied in my math degree - other texts were not quite as transparent to me until I learned a bit about Rigged Hilbert Spaces - but that is another issue) I seem to recall that's pretty much what he did.

Wigner abandoned it when he heard of some early work on decoherence by Zurek.

Fredrik said:
I think the idea that the basis is independent of the decomposition is as likely to be true as the idea that 2x is independent of x.

And you may be right - however I prefer not to be so sure about it until the theorems demonstrating it unequivocally one way or another are forthcoming and they have had time to be checked. It is interesting standard texts on dechorenece such as the one I have make no mention of it - there may be something already known about it we are missing - I am adopting a wait and see attitude to it..

Thanks
Bill
 
Last edited:

Similar threads

Replies
130
Views
9K
Replies
29
Views
5K
Replies
69
Views
6K
Replies
14
Views
2K
Replies
28
Views
3K
Back
Top