Why MWI cannot explain the Born rule

Click For Summary
The discussion centers on the argument that the Many-Worlds Interpretation (MWI) of quantum mechanics cannot adequately explain the Born rule using its minimal assumptions. It posits that if MWI's assumptions lead to a probabilistic interpretation, then any system adhering to these must conform to the Born rule. However, counterexamples from classical mechanics demonstrate systems that meet these criteria without yielding a probabilistic interpretation. Participants debate the implications of defining "worlds" as correlations and the necessity of additional axioms to incorporate the Born rule effectively. Ultimately, the conversation highlights the challenge of reconciling deterministic theories with probabilistic outcomes in quantum mechanics.
  • #121
Demystifier said:
Second, a measurement can be described even by describing the whole system. It may be impossible in practice due to a large number of the degrees of freedom, but it is possible in principle.
I am under the impression that is actually a significant point of contention between interpretations -- e.g. that Copenhagen says it's impossible in principle.

(unless, of course, you switch to a new and better physical theory)
 
Physics news on Phys.org
  • #122
Fredrik said:
I wonder if it's possible to prove that the Born rule is the only probability measure that gives us P(a,b)=P(a)P(b).

I wonder if you like this, more than I do?

Ariel Caticha's
Consistency, Amplitudes and Probabilities in Quantum Theory "Quantum theory is formulated as the only consistent way to manipulate probability amplitudes. The crucial ingredient is a consistency constraint: if there are two different ways to compute an amplitude the two answers must agree. This constraint is expressed in the form of functional equations the solution of which leads to the usual sum and product rules for amplitudes. A consequence is that the Schrödinger equation must be linear: non-linear variants of quantum mechanics are inconsistent. The physical interpretation of the theory is given in terms of a single natural rule. This rule, which does not itself involve probabilities, is used to obtain a proof of Born’s statistical postulate. Thus, consistency leads to indeterminism."
-- http://arxiv.org/PS_cache/quant-ph/pdf/9804/9804012v2.pdf

In effect I think he is trying to generalize Cox, Jaynes "derivation" of the rules of koglomorov's probability, not from the traditional axioms, but from some assumptions of consistency of reasoning around information.

One key assumption is that he assumes as part of the microstate of information beeing represented/quantified by a REAL numbers (or degree of beliefs). Then by a series of arguments he shows that the only consistent logical system och ang or operators etc that fulfills this (and some hidden mor or less natural assumptions) is indistinguishable from koglomorov probability axioms.

The generalisation is to isntead, assume (again key assumption) that instead the state of information is represented/quantified by a COMPLEX number. Then his idea is that similarly quantum logic emerges as the only consistent system. He argument for the Born rule is I think effectively the same as the ones trying to "count" distinguishable microstates and arrive at some kind of "frequency".

I think both of these ideas as intersting but my objection to both of those are the key assumptions of choose real or complex numbers as representation.

Ariel points out this himself:
"A second, simpler question is why do we seek a representation in terms of complex numbers? Again, no answer here; this is an unexplained feature of quantum theory. It seems that a single complex number is sufficient to convey the physically relevant information about a setup."

So I think there mere starting point of a normed linear vector space over complex numbers is quite non-trivial as well.

For me, the physical insight of the meaning of real numbers in a framework where only a finite amount of information is encodable is still mysterious. Before I understand this, I can not claim to be content with any higher level derivation which contains this as unmotivated baggage.

I'm curious if Fredrik, finds Ariels paper more satsifactory than I do?

/Fredrik
 
  • #123
Fredrik said:
That's right (and also obvious). I just proved that if we use the tensor product and the Born rule, we get P(a,b)=P(a)P(b) for non-interacting systems. (It would be a disaster to get P(a,b)≠P(a)P(b). If QM works for nuclei and electrons separately, it wouldn't work for atoms. But of course it wouldn't work for nuclei either...) The point is that quantum mechanics for individual systems, which by definition includes the Born rule, more or less forces us to us to use the tensor product to represent the states of the composite system.

Do you picture that the notion of non-interacting systes still makes sense when you incorporate say gravity? how do you prevent the gravitational interaction?

/Fredrik
 
  • #124
Fra said:
I wonder if you like this, more than I do?
...
I'm curious if Fredrik, finds Ariels paper more satsifactory than I do?
My impression (after spending about half an hour on it) is that he seems to have found a way to understand probability amplitudes on a slightly deeper level. It looks good, but not very significant.

Fra said:
Do you picture that the notion of non-interacting systes still makes sense when you incorporate say gravity? how do you prevent the gravitational interaction?
I'm not going to try to unify QM with GR today. Maybe tomorrow. :smile:
 
  • #125
dmtr said:
Well, my advice to these inhabitants would be "use the symmetry and the number of simulation copies to derive the probabilities for future events". Following this advice will help these inhabitants to predict the future better.

You're still missing the point. How do you know what probability to use in a simulation if empirically all you ever have access to is one branch of the outcome? He's giving you a God's eye view so you can see that it's impossible for the individual inhabitants to do science. We're an individual inhabitant, not God.
 
  • #126
Demystifier said:
I don't think that it is correct.

First, no physical process turns pure states into mixed states, provided that the whole system is taken into account. Such a process would contradict unitarity.
I meant approximately. I was just too lazy to explain that (and I thought people would understand what I meant after I had just explained it for the first type of process).
 
Last edited:
  • #127
Fredrik said:
I'm not going to try to unify QM with GR today. Maybe tomorrow. :smile:

I sure won't do that neither today nor tomorrow either :)

But I guess my attitude is that I see indications that the deeper understanding of the foundations of QM, and a possible deeper motivation for QM and it's structure, might be additionally complicated by us trying to keep gravity out of it. Although on the surface foundational QG for sure looks more hairy than foundational QM, trying to find a connection as deep as possible might also enlighten us with regards to normal QM, and the structure of the standard model of particle physics in the quest for a GUT. Maybe gravity is simply the "missing link". Maybe it's a mistake to think that trying to think of gravity already from start will make the quest harder, rather than easier?

/Fredrik
 
  • #128
Fredrik said:
I'm not going to try to unify QM with GR today. Maybe tomorrow. :smile:

I'm hoping tomorrow ... or the day after. I have the equation, I just need a solution to get the party started :smile:
 
  • #129
Fra said:
Maybe it's a mistake to think that trying to think of gravity already from start will make the quest harder, rather than easier?
That's certainly possible. The difference between non-relativistic QM and special relativistic QM is a different group of symmetries for the theory (or equivalently, a different algebra of observables). Is there a difference between special relativistic QM and general relativistic QM? Probably. I think that's what LQG is trying to answer. I also think that almost everything that's been written about attempts to interpret QM or its underlying mathematical structure as a description of what actually happens, have completely ignored those differences.
 
  • #130
jensa said:
I have been meaning to ask a question related to some of the issues of this thread namely; Does decoherence require a an environment (decomposition into subsystems)?
It think the answer is yes. I don't know decoherence well, but the impression I got is that the crucial step is the calculation of a reduced density matrix, as discussed above. This isn't possible without a decomposition into subsystems. If you want a better answer, try searching arxiv.org for articles by Zurek, or buy Schlosshauer's book. (I haven't read it yet, but it's in my shopping cart).
 
  • #131
Fredrik said:
It think the answer is yes. I don't know decoherence well, but the impression I got is that the crucial step is the calculation of a reduced density matrix, as discussed above. This isn't possible without a decomposition into subsystems. If you want a better answer, try searching arxiv.org for articles by Zurek, or buy Schlosshauer's book. (I haven't read it yet, but it's in my shopping cart).

Thank you for the response Fredrik,

I don't think that decomposition into subsystems is sufficient (maybe not even necessary) to produce decoherence. Consider two sets of interacting two-level systems. The Hilbert space of this composite system we describe as a tensor product of the individual Hilbert spaces. If we are only interested in the properties (observables) of one of the subsystems we may trace out the other one to produce a reduced density matrix. The reduced density matrix will generally be a mixed state but the coherence factors (off diagonal elements of the density matrix) need not vanish irreversibly. Most likely you would observe an oscillatory behaviour at a certain frequency. To observe irreversible behaviour you need to also assume that the system you are tracing out contains a large (infinite) number of degrees of freedom so that different frequencies add up to produce a decay on average.

It seems to me that the macroscopic nature of the environment (many degrees of freedom) is more important than the decomposition into subsystems in order to observe irreversible loss of coherence.

I hope I don't project too much of my own ignorance onto the people on this board but loss of coherence seems to be similar to the classical increase of entropy in that everybody believes it to occur but very few can actually show it (and explain what causes it). In the case of increase of classical entropy there seems to be a number of ways how to justify it. Personally I prefer the coarse grained explanation that the increase of our ignorance (entropy) comes from our mapping from the microscopic configurations towards macroscopic observables. I.e. it can be shown that when we map the many-particle phase space onto a space of macroscopic observables, and replace the distribution function \rho(X,t) where X is the coordinate in phase space, by the distribution function \tilde{\rho}(A,t) where A is a coordinate in the space of macroscopic observables, the entropy defined in terms of the macroscopic distribution function \tilde{\rho} always increases.

I believe the origin of decoherence has a similar source, namely that it is the mapping from microscopically distinguishable states onto macroscopically distinguishable observables that produces a practical impossibility to observe interference effects of macroscopic objects. As in my example with Schrödingers cat; our ignorance about the microscopic states along with the extremely short time scales of microscopic processes causes the interference effects to become "averaged out".

Of course, when we are talking about the decoherence of truly microscopic systems one needs to consider the entanglement with a macroscopic environment. Now the observables we are interested in are of course the observables associated with the original microscopic system (so they are not macroscopic observables). But still there exists a many-to-one mapping corresponding to our ignorance about the many degrees of freedom of the environment.

I wish I could be more thorough with my explanations but I have a lot of work to do. Btw, how do you guys manage to spend so much time on the boards and learn a lot of new stuff?

PS. I have Schlosshauers book and will try to read it when I get some time.
 
  • #132
RUTA said:
I'm hoping tomorrow ... or the day after. I have the equation, I just need a solution to get the party started :smile:

What was your general idea? (There was probably a past thread about this, but I can't remember)

/Fredrik
 
  • #133
Hurkyl said:
I am under the impression that is actually a significant point of contention between interpretations -- e.g. that Copenhagen says it's impossible in principle.
I guess you have in mind the idea that not everything can be described by QM, but that you also need a classical world. I would say that this is only one of several different variants of the Copenhagen interpretation. And I think that this particular variant is quite obsolete.
 
  • #134
Some of you mentioned the Schlosshauer book. Let me just say that I recommend it to everyone. Today no one can say that he understands QM well without being familiar with basics of decoherence. Actually, this book contains more than just basics, but it is worth reading.
 
  • #135
Demystifier said:
I would say that this is only one of several different variants of the Copenhagen interpretation. And I think that this particular variant is quite obsolete.
My impression is that CI always has the property that collapse is real (as opposed to a mathematical technique or a change-of-frame-type thing). I know of three variants:
  1. Quantum and classical mechanics tell us what's really going on1, but QM becomes inaccurate above a certain scale and CM becomes inaccurate below a certain scale. We can effectively use the two in concert by invoking a Heisenberg cut.
  2. QM doesn't tell us anything about what's really going on -- it is a theory of our ignorance2 of deeper reality
  3. QM will work all the way up, once we figure out what nonlinear terms should be inserted into Schrödinger's equation to make collapse occur

(note this last one still asserts that unitary evolution of states is wrong on macroscopic scales)

Is there another variant you had in mind?


1: Meaning, roughly, that the elements of the theory correspond to elements of reality
2: I don't mean this pejoratively -- I mean it as in "ignorance probabilities"[/size]
 
  • #136
Hurkyl said:
Demystifier said:
Second, a measurement can be described even by describing the whole system. It may be impossible in practice due to a large number of the degrees of freedom, but it is possible in principle.

I am under the impression that is actually a significant point of contention between interpretations -- e.g. that Copenhagen says it's impossible in principle.

(unless, of course, you switch to a new and better physical theory)

There was a recent discussion about the "definition of CI". But to leave aside that classification of views here, the reason why I think it's impossible even in principle is this:

As I see it, a given observer doesn't in general "choose" to observe this or that. The "complete" picture (as I see it) is that a given observer always "observes" (ineracts with) it's own environment.

But no given observer, can relate to, and decode all possible the degrees of freedom in the environment. So each observer sees a truncated world.

Now, I take a similar to rovelli's RQM view here, that the only way to level two "truncated worlds" against each other, is by means of the observers interacting. The view of each observer, will then be revealed by the way that act upon each other.

Like a game of poker. Each player has "a vision" of the future of the game - if my opponent folds, I have good grounds to think it's because he thinks he has slim chance to win - his vision is revealed to me. The only way to find out the opponents "visions" are to play the game.

The problem in QM, is that even though it's true that a second observer (possibly a massive one) COULD in some approximation observe the measurement process in the sense of "environment as the observer", this massive amount of information could never be conveyed back to the original observer - even if a perfect communication channel was established - simply because they are not comparable in complexity.

The analogy to say SR or GR would be perfect if we could establish the transformations that restore observer invariance here, and view these transformations in the realist sense.

But one problem is that there is no way that there can exist one-2-one transformations between structures that can't not encode the same amount of information. The transformations themselves must create and destroy information unless we have an *equilibrium situation* where the environment contains copies of the same repeating and thus "redundant" information, then we find the special case that a truncated system might contain the same information as a larger system.

But if we consider the general non-equilibrium case, I think the information preserving idea simply won't do.

I think that to understand QM, would be to find the general case, and then see why and how QM structure as we know it, does emerge as a uniqe special case. Like GR vs SR.

/Fredrik
 
  • #137
There have been several discussions about the CI recently. There was a thread started by Demystifier, and this thread started by me. There are links to interesting papers in #7 and #20.

My impression is that the original CI is essentially the same as what we've been calling "the ensemble interpretation" in this forum. There's a formal difference in their definitions, but I don't see how that formal difference is an actual difference. See #33 in the thread I linked to above. Hurkyl, your #2 appears to be the same thing, but if QM doesn't tell us what actually happens, then there's nothing that suggests that "collapse" is a physical process, as you suggested at the start.
 
  • #138
Fra said:
What was your general idea?

A discrete path integral over graphs. The difference matrix K and source vector J are constructed from boundary operators in the spacetime chain complex of the graph so that Kx = J where x is the vector of nodes, links or plaquettes in the resulting scalar, vector or tensor field theory, respectively (this follows from the boundary of a boundary principle, BBP). This restricts K and J in the discrete action of the transition amplitude, Z. Note that Z is not a function of the field Q, i.e., Q is the integration variable in computing Z, yet quantum and classical field theories are all about Q. How is that?

When you ask for the probability that the kth node, link or plaquette has the value Qo you obtain Z(Qk=Qo)/Z (Z is a partition function since we're using a Euclidean path integral), which is the discrete counterpart to QFT. If you ask for the most probable value of Qo, you find you must solve KQo = J, i.e., the discrete counterpart to CFT. Notice that by requiring the graphical basis satisfies the BBP, your discrete CFT automatically satisfies the BBP, which is the basis for the local conservation of momentum and energy (divergence-free stress-energy tensor). These are standard calculations, we're just proposing a different take on them that leads to a discrete CFT at odds with GR.

Anyway, we're trying to solve the tensor version of KQo = J for some simple situation and compare the result to GR to "get the party started." We expect differences from GR since our version of discrete tensor CFT is linear and constitutively non-local while GR is non-linear and local. Essentially, our discrete tensor CFT is Regge calculus where the nodes of the simplices are clusters of graphical elements (think of clusters of cubes joined by line segments) so there are no vacuum solutions (it's constitutively non-local aka inseparable -- no empty spacetime, just spacetimematter) and the lengths of the line segments joining the clusters are simply equal to the average values of the Q's on the shared cube faces (plaquettes) between clusters (average value = most probable value since we've a Gaussian distribution function). This resolves the problem resulting from the fact that violations of Bell's inequality imply causal and/or constitutive non-locality while GR is local on both counts, in favor of QM (GR must be revised).

We have a paper under review at Foundations of Physics. If that gets accepted and we find our classical solution, then maybe I'll start a thread :smile: We can't discuss this anymore here, it's out of context.
 
  • #139
I have now read enough of this paper to see that the fact that we should be using the tensor product to represent a composite system can be derived without direct reference to the Born rule. It's based on the quantum logic approach to QM, which associates a mathematical structure with the set of statements of the form "if you measure observable A, you will get a result in the set B with probability 1". So it has some connection to probabilities, but it's not as strong as I expected. Maybe the stuff I mentioned above about how we need to use the tensor product to ensure that the Born rule satisfies P(a,b)=P(a)P(b) for non-interacting systems is in there somewhere, and I just can't see it.
 
  • #140
What about http://arxiv.org/abs/0903.5082 which tries to derive the Born rule from

(i) States are represented by vectors in Hilbert space
(ii) Evolutions are unitary
(iii) Immediate repetition of a measurement yields the same outcome

"To derive it we cannot use reduced density matrices, Eqs. (1,2). Tracing out is averaging [25, 29, 30] - it relies on pk = |psik|2, Born's rule we want to derive."
 
Last edited:
  • #141
RUTA said:
A discrete path integral over graphs.
...
When you ask for the probability that the kth node, link or plaquette has the value Qo you obtain Z(Qk=Qo)/Z (Z is a partition function since we're using a Euclidean path integral), which is the discrete counterpart to QFT. I
...
We have a paper under review at Foundations of Physics. If that gets accepted and we find our classical solution, then maybe I'll start a thread :smile: We can't discuss this anymore here, it's out of context.

Thanks for the hints, just a couple of quick questions :)

I got the impression from other threads you are seeking a reconstruction of the continuum in terms of a discrete model that is more "fundamental"? I symphatise because I also find the continuum starting point inherently unphysical uncountable redundance that isn't helping at all.

Does this mean you also reconstruct a "discrete" probability theory, where the measure rather spans a discrete rational subset of [0,1], constrained by complexity?

Somehow this latter thing, is I think relevant to the discussion of defining probability since it allows to avoid the issue of "infinite measurements" and frequency limits. Instead finite information might imply that the probability measure itself is discretized and not covering a continuum. This would suggest that one could actually "count" the truncated contiuum and also define measures on stuff like the space of possibilities (which is used for feynmann summation).

/Fredrik
 
  • #142
Fra said:
Thanks for the hints, just a couple of quick questions :)

I got the impression from other threads you are seeking a reconstruction of the continuum in terms of a discrete model that is more "fundamental"? I symphatise because I also find the continuum starting point inherently unphysical uncountable redundance that isn't helping at all.

Does this mean you also reconstruct a "discrete" probability theory, where the measure rather spans a discrete rational subset of [0,1], constrained by complexity?

Somehow this latter thing, is I think relevant to the discussion of defining probability since it allows to avoid the issue of "infinite measurements" and frequency limits. Instead finite information might imply that the probability measure itself is discretized and not covering a continuum. This would suggest that one could actually "count" the truncated contiuum and also define measures on stuff like the space of possibilities (which is used for feynmann summation).
Yes, the discrete structure is fundamental to the continuum structure, not a mere approximation thereto.

Our partition function (transition amplitude, Z) is defined over a countable number of graphical elements, but each element can have an uncountable number of possible field values.
 
  • #143
jensa said:
I don't think that decomposition into subsystems is sufficient (maybe not even necessary) to produce decoherence. Consider two sets of interacting two-level systems. The Hilbert space of this composite system we describe as a tensor product of the individual Hilbert spaces. If we are only interested in the properties (observables) of one of the subsystems we may trace out the other one to produce a reduced density matrix. The reduced density matrix will generally be a mixed state but the coherence factors (off diagonal elements of the density matrix) need not vanish irreversibly. Most likely you would observe an oscillatory behaviour at a certain frequency. To observe irreversible behaviour you need to also assume that the system you are tracing out contains a large (infinite) number of degrees of freedom so that different frequencies add up to produce a decay on average.
Sounds like you know a few things about decoherence that I don't. But you specifically mention reduced density matrices, and those can't even be defined without a tensor product decomposition. I have started reading a pdf version of Schlosshauer (I'm still buying the real one) and in the intro, he describes decoherence as the system getting more and more entangled with the environment. Everything I have seen indicates that you need to consider at least two component subsystems: "the system" and "the environment".
 
  • #144
atyy said:
What about http://arxiv.org/abs/0903.5082 which tries to derive the Born rule from...
Thanks. I intend to check it out, but I'll probably wait until I've read some more in Schlosshauer's book. I expect that I will have objections about the use of the tensor product and the use of density matrices. The former has some connection with probabilities that I don't fully understand yet (see #139), and the latter seems impossible to justify without the Born rule (see #108).
 
  • #145
Fredrik said:
But you specifically mention reduced density matrices, and those can't even be defined without a tensor product decomposition.

Yes of course you are right, but my point with the passage you quoted was that even if we can decompose a system into subsystems (with tensor product) and look at the reduced density matrix of the particular subsystem we are interested in, you still do not necessarily get irreversible loss of coherence! Sure, the reduced density matrix will generally be of the mixed kind but will it irreversibly go towards a total loss of coherence?

This depends on what the other subsystem is. Mostly when people talk about "environment" I believe it is implied that it consists of an infinite amount of degrees of freedom (i.e. it is macroscopic). In other words, the macroscopic nature is at least as important to the concept of decoherence as the decomposition into subsystems. My general point is that you can actually remove the decomposition as a necessity.

Let me try one last time to convince you that you can get a loss of coherence even without the decomposition feature by returning to my original example of Schrödingers cat.

Let us write the general microscopic state of the macroscopic system (cat) as:

<br /> |\psi\rangle = \sum_i c_i(t)|i\rangle<br />

where i here denotes a set of labels completely characterizing the microscopic state of all particles the cat consists of (clearly a huge amount). Let us assume that we have chosen a basis in such a way that we can clearly distinguish for which microscopic state |i\rangle the cat is either dead or alive. We can formally define a set of projection operators:

<br /> \hat{P}_\text{alive}=\sum_{i\in alive}|i\rangle\langle i|, \quad \hat{P}_\text{dead}=\sum_{i\in dead}|i\rangle\langle i|<br />

We can then associate the state \hat{P}_\text{alive}|\psi\rangle=\sum_{i\in \text{alive}}c_i(t)|i\rangle with a macroscopic state of the cat being alive and vice versa with the macroscopic "dead state". If we, for a moment, trust the conventional probability rule we have:

<br /> \text{Prob. alive}=\langle \psi|\hat{P}_\text{alive}|\psi\rangle=\sum_{i\in \text{alive}}|c_i(t)|^2, \quad \text{Prob. dead}=\langle \psi|\hat{P}_\text{dead}|\psi\rangle=\sum_{i\in \text{dead}}|c_i(t)|^2<br />

So far so good, but is it possible to define a "relative phase" between the two macroscopic states "dead" and "alive"? In principle it should be clear already here that such a feat is difficult and designing an interference experiment without knowing the microscopic configuration of the cat is pretty much impossible. However, we could analyze the operator

<br /> \hat{P}_\text{a-d}=\sum_{i\in \text{alive}}\sum_{j\in \text{dead}}|i\rangle\langle j|<br />

which connects the dead and alive subspaces and thus the object

<br /> \langle \psi|\hat{P}_\text{a-d}|\psi\rangle=\sum_{i\in \text{alive}}\sum_{j\in \text{dead}}c_i^*(t)c_j(t)<br />

is directly related to the "coherence". Now there are two issues here: 1) The c_i's are determined by the exact microscopic state (determined by initial conditions and exact many-particle hamiltonian) of which we are clearly ignorant. 2) the time scale of variation of the c_i(t) is very short compared to macroscopic time scales. We thus expect this object to fluctuate wildly (both statistically and temporally) i.e. it is essentially a chaotic variable. Performing an average (coarse graining) over this object will average it out. The probabilities on the other hand must of course add up to unity and, while it may fluctuate depending on initial conditions and as a consequence of time dependence of c_i(t), because it is always a positive quantity between 0 and 1 it will average to some constant. What this is supposed to illustrate is that the off-diagonal elements will vanish upon statistical and temporal averaging, and it is effectively impossible to create an interference experiment that may be used to observe the relative phase between the macroscopic states "dead" and "alive".

I have started reading a pdf version of Schlosshauer (I'm still buying the real one) and in the intro, he describes decoherence as the system getting more and more entangled with the environment. Everything I have seen indicates that you need to consider at least two component subsystems: "the system" and "the environment".

Yes he seems to credit the non-locality of quantum mechanics, which of course is related to system+environment. In fact most textbooks seem to use this notion to describe decoherence. However, I feel like this is only one specific type of decoherence (environmentally induced decoherence) and that decoherence in general can be described without the use of an external environment.EDIT: Sorry everyone for the long post..hope I didn't derail the discussion too much.

/Jens
 
Last edited:
  • #146
Count Iblis said:
The argument by Hartle allows you to replace the Born rule by the weaker rule that says that measuring an observable of a system if the system is in an eigenstate of that observable, will yield the corresponding eigenvalue with certainty.
I've been doing some more thinking about this. I still think that Hartle's argument is useless, and proves nothing, but I've realized that Gleason's theorem says something very similar to the above. It says that if \mu is a probability measure on the set of subspaces on a separable Hilbert space (real or complex, and at least 3-dimensional), there exists a density operator \rho such that

\mu(M)=\mbox{Tr}(\rho P_M)

where P_M is the projection operator associated with the closed subspace M.

Consider the simplest possible case, i.e. when \rho is a pure state |\psi\rangle\langle\psi|, and M is a 1-dimensional eigenspace corresponding to the eigenvalue b of an observable B. The theorem says that the only possible probability measure assigns probabilty

\mbox{Tr}(|\psi\rangle\langle\psi|b\rangle\langle b|)=\sum_{b&#039;}\langle b&#039;|\psi\rangle\langle\psi|b\rangle\langle b|b&#039;\rangle=|\langle b|\psi\rangle|^2

to that eigenspace. So it certainly looks like Gleason has derived the Born rule. There are however several subtle points worth noting here.

1. Why are we looking for probability measures on the set of closed subspaces of a separable Hilbert space? A partial answer is that the closed subspaces can be thought of as representing "properties" of physical systems in QM*. This is an axiom in the quantum logic approach to QM**, but in the traditional Hilbert space approach to QM, we would have to use the Born rule to prove that this is true.

2. Where did we "put probabilities into get probabilities out"? This is explained by item 1 and the footnotes. The probability measure (which is uniquely determined by the state) assigns non-trivial probabilities to mathematical objects (closed subspaces) that are already associated with assignments of probability 1 to possible events in the real word. (This is where it's very similar to what Count Iblis claimed above).

3. This is clearly not a derivation of the sort originally envisioned by Everett. It isn't a derivation from the assumption that the state of the universe can be described by a state vector satisfying a Schrödinger equation. Gleason didn't assume that. Instead he started with the assumption that the set of "properties" is represented by the simplest possible mathematical structure that's consistent with a set of axioms that we expect all theories to satisfy. Of course, before QM was discovered, we would have guessed that all theories must satisfy a much stronger set of axioms, so this set of axioms was chosen specifically to ensure that QM (with its Born rule) qualifies as a theory.

*) What I call "properties" goes by many names in the literature, including "propositions", "elements of reality" and "experimentally verifiable statements". The last one is probably the most appropriate, since these phrases all refer to the possible results of experiments that are assigned probability 1 by QM.

**) Technically they use another set of axioms in order to associate a mathematical structure with the set of properties, but then they define a "standard" structure as one that's isomorphic to the lattice of closed subspaces of a complex separable Hilbert space
 
Last edited:
  • #147
Just to conclude my somewhat tangential series of posts: What I have been describing falls under a category which Joos calls "Fake Decoherence"*, while he prefers to restrict the word decoherence to the system+environment stuff. Personally I find the term "fake decoherence" quite misleading; the effect of practical inability to observe coherence is quite real. I would prefer to call the effect itself decoherence and then use "environmentally induced decoherence" to refer to the system+environment stuff. Of course it is quite possible that for the purposes of MWI only environmentally induced decoherence is important, although I cannot see immediately why this would be so.

*) See the book "Decoherence and the appearance of a classical world in quantum theory" by Joos, Zeh, Kiefer, Giulini, Kupsch and Stamatescu in the section entitled "True, False and Fake decoherence". I personally prefer this book over Schlosshauer's.
 
  • #148
jensa said:
*) See the book "Decoherence and the appearance of a classical world in quantum theory" by Joos, Zeh, Kiefer, Giulini, Kupsch and Stamatescu in the section entitled "True, False and Fake decoherence". I personally prefer this book over Schlosshauer's.
This is an excellent book too. However, I cannot find the section you mention above. Can you help me (section number, page number, contributor name, or something like that)?
 
  • #149
Demystifier said:
This is an excellent book too. However, I cannot find the section you mention above. Can you help me (section number, page number, contributor name, or something like that)?

I have the second edition where it is in Chapter 3 "Decoherence Through interaction with the Environment" by Joos, section 3.4.3. Perhaps it is absent in the first edition?

Edit: You can find the second edition by searching for it in Google Books.
 
  • #150
jensa said:
I have the second edition where it is in Chapter 3 "Decoherence Through interaction with the Environment" by Joos, section 3.4.3. Perhaps it is absent in the first edition?

Edit: You can find the second edition by searching for it in Google Books.
Yes, I have the first edition. It does not even contain Sec. 3.4.3.

The preview by Google Books gives first 101 pages, which does not cover Sec. 3.4.3 either.
 
Last edited:

Similar threads

  • · Replies 47 ·
2
Replies
47
Views
6K
  • · Replies 8 ·
Replies
8
Views
3K
  • · Replies 27 ·
Replies
27
Views
3K
  • · Replies 60 ·
3
Replies
60
Views
9K
  • · Replies 11 ·
Replies
11
Views
3K
  • · Replies 11 ·
Replies
11
Views
3K
  • · Replies 34 ·
2
Replies
34
Views
6K
  • · Replies 309 ·
11
Replies
309
Views
16K
  • · Replies 76 ·
3
Replies
76
Views
8K
  • · Replies 2 ·
Replies
2
Views
2K