Can the Born Rule Be Derived Within the Everett Interpretation?

  • Thread starter Thread starter vanesch
  • Start date Start date
  • Tags Tags
    Born rule Paper
  • #101
straycat said:
Where does Zurek make this assumption -- implicitly or otherwise?

He does it implicitly, in two places. He first does it when he introduces the states |C_k> in equation 8b, and his hilbert space HC of sufficient dimensionality in 9a. Clearly, he's now supposing a fine-grained measurement, where the c_j states are measured too, and from which are DERIVED afterwards the probabilities for the eventual coarse grained measurement. As such, he implicitly assumes that the the coarse grained measurement will give you the SAME probabilities as the sums of the probabilities of the fine grained measurement.

But he KNOWS that he's doing something fishy ! On p18, he writes (just under 1. Additivity...
In the axiomatic formulation ... as well as in the proof of the Born rule due to Gleason, additivity is an assumption motivated by the mathematical demand...

And he tries to weasel out with his Lemma 5 and his probabilities calculated from the state (27) "representing both fine-grained and coarse-grained records". However, he effectively only considers the probabilities of the fine-grained events.

Again, we will use our non non-contextual example to illustrate the flaw in his proof:

we consider |psi> = |x1>|y1> + |x1>|y2> + |x2>|y3>

As such, for the (finegrained) Y measurement, we have:
P_f(y1) = 1/3
P_f(y2) = 1/3
P_f(y3) = 1/3

and thus: P_f(x1) = 2/3 and P_f(x2) = 1/3

However, for the coarsegrained X measurement, we have:
P_c(x1) = 1/2
P_c(x2) = 1/2

AND IT MAKES NO SENSE TO TALK ABOUT THE PROBABILITY OF THE FINEGRAINED EVENTS. If I were to talk about the probabilities of y1, y2 and y3 for the probability measure P_c, I would get nonsense of course.

From the moment you ASSIGN a probability to the finegrained events, of course from the Kolmogorov axioms, additivity is implicitly incorporated.

Only, Zurek uses only ONE probability function, p(). As he is considering probabilities of fine-grained events in his subtraction procedure, the p() is the finegrained probability measure (P_f in my example). There of course, additivity is correct.

He's assuming that the probability function is the SAME ONE for fine grained and coarse grained measurements, and that is nothing else but the (rightly identified) extra assumption of Gleason of non-contextuality. But he's making the same assumption in his Lemma 5!
 
Physics news on Phys.org
  • #102
straycat said:
Hmm. It seems to me that assuming additivity of probabilities is fine, if you assume that (to use my example above) the spin measurement is in fact the more fine-grained measurement of the exact value of n. I suppose n could be called a "hidden variable," and when we think we are only measuring spin, we are in fact measuring this hidden variable -- we just haven't been smart enough to figure it out yet!

Ok, but in that way, I can produce you ANY probability measure that is compatible with unitary dynamics: the APP, the Born rule, any other function that does the trick. If I'm allowed to say that the measurement of an observable O1 is in fact the measurement of the observable O1 x O2, where O2 works onto a yet to be specified Hilbert space with a yet to be established number of degrees of freedom and a yet to be established dynamics (interacting with O1) so that I get out the right number of "different" outcomes, I can provide you with just ANY probability rule.

But even there, you have a problem when I change something. Suppose that I start from a state u1|a> + u2|b> and I do a binary measurement (a versus b). Now, you claim that there is some physics that will evolve:

|a> (|x1>+|x2> +...|xn>) + |b> (|y1> + ... ym>)

such that n is proportional to u1^2 and m is proportional to u2^2, and that my "binary measurement" is in fact a measurement of the x1... ym states. Ok.

But suppose now that I'm measuring not u1|a> + u2|b>, but rather u2 |a> + u1 |b>. If we have the same unitary evolution of the measurement, I would now measure in fact the x1... ym states in the state:

u2/u1 |a> (|x1>+|x2> +...|xn>) + u1/u2 |b> (|y1> + ... ym>)

right ?

But using the APP, I would find probability |u1|^2 for |a> and |u2|^2 for |b> and not the opposite, no ?

Why would the dimensionality of the x1...xn depend on the coefficient u1 of |a> in the original state ? This cannot be achieved with a unitary
operator which is TRANSPARENT to the coefficient.

Isn't this a fundamental problem to assuming a certain dimensionality of hidden variables in order to restore the Born rule ?
 
  • #103
vanesch said:
Ok, but in that way, I can produce you ANY probability measure that is compatible with unitary dynamics: the APP, the Born rule, any other function that does the trick. If I'm allowed to say that the measurement of an observable O1 is in fact the measurement of the observable O1 x O2, where O2 works onto a yet to be specified Hilbert space with a yet to be established number of degrees of freedom and a yet to be established dynamics (interacting with O1) so that I get out the right number of "different" outcomes, I can provide you with just ANY probability rule.

Yes, you are correct: You could in fact take Zurek's basic idea and come up with any probability rule! This is what I mean by "baby steps." I do believe that Zurek has avoided the circularity trap. What he has not done afaict is to demonstrate why the Born rule, and not some other rule, must emerge. But that is progress, no?

So now we turn to your next argument:

vanesch said:
Why would the dimensionality of the x1...xn depend on the coefficient u1 of |a> in the original state ?

I believe that some additional rule or set of rules is necessary to answer this question. And the sole motivation for postulating the "Born constraints" in my draft manuscript is to provide an "existence proof" that it is possible to accomplish this.

vanesch said:
This cannot be achieved with a unitary
operator which is TRANSPARENT to the coefficient.
Isn't this a fundamental problem to assuming a certain dimensionality of hidden variables in order to restore the Born rule ?

I'm not sure I entirely follow your argument that this cannot be achieved. I have a feeling, though, that the answer has something to do with the fact that you need to consider, not only the state of the system under observation, but also the state of the measurement apparatus. To use your example above,

|a> (|x1>+|x2> +...|xn>) + |b> (|y1> + ... ym>)

suppose the "binary measurement" is a spin measurement along the x-axis. We could suppose that the number of dimensions of the fine-grained measurement has something to do with the interaction between the particle and the SG apparatus. IOW, if the SG apparatus is oriented to measure along the x-axis, then the relevant "number of dimensions" is n and m (following your notation above). But if we rotate the SG apparatus so that it measures along some other axis, then the relevant number of dimensions becomes n' and m'. Of course, it still is necessary to explain WHY this should work out just right, so that the Born rule emerges. But the point I wish to make is that there is no reason to ASSUME that this CANNOT be done! Unless I have missed some element of your argument, which is why I am enjoying this discussion ...:biggrin:

David
 
  • #104
vanesch said:
... we consider |psi> = |x1>|y1> + |x1>|y2> + |x2>|y3>
As such, for the (finegrained) Y measurement, we have:

P_f(y1) = 1/3
P_f(y2) = 1/3
P_f(y3) = 1/3

and thus: P_f(x1) = 2/3 and P_f(x2) = 1/3

However, for the coarsegrained X measurement, we have:

P_c(x1) = 1/2
P_c(x2) = 1/2

AND IT MAKES NO SENSE TO TALK ABOUT THE PROBABILITY OF THE FINEGRAINED EVENTS. If I were to talk about the probabilities of y1, y2 and y3 for the probability measure P_c, I would get nonsense of course.

You raise the issue: given a measurement of the above system, which should we use: P_f or P_c? How do we justify using one and not the other?

Following the spirit of Everett's original proposal, I believe that the number of branches (ie, the "number of dimensions") associated with a given measurment must be reflected in the number of distinct physical states that the observer can evolve into, as a result of the measurement process. So if the interaction of the observer with the environment results in evolution of the observer from one to 3 different possible states, then we have 1/3 probability associated with each state. If two of these observer-states are associated with x1, and the third is associated with x2, then we get:

P_c(x1) = 2/3
P_c(x2) = 1/3

So the above result, yielding probabilities 2/3 and 1/3, depends upon the assertion that there are two mutually exclusive distinct physical observer-states associated with x1, but only one observer-state associated with x2. A fully developed underlying theory must give an exact prescription for this number of observer states, as well as tell us which states are associated with which observable (x1 or x2).

My point is that the choice between P_f and P_c is not arbitrary, but should be uniquely determined by the underlying theory, which must (if it is going to work) describe the evolution of the physical state of the observer. This underlying theory has not been found yet, but I think it will be!

David
 
  • #105
straycat said:
Yes, you are correct: You could in fact take Zurek's basic idea and come up with any probability rule! This is what I mean by "baby steps." I do believe that Zurek has avoided the circularity trap. What he has not done afaict is to demonstrate why the Born rule, and not some other rule, must emerge. But that is progress, no?

Eh ? What progress ? That we can have any probability rule ? :-p

So now we turn to your next argument:
I believe that some additional rule or set of rules is necessary to answer this question.

:biggrin: :biggrin: :biggrin:

That's what I'm claiming all along! Now why can that extra rule not simply be: "use the Born rule" ?

And the sole motivation for postulating the "Born constraints" in my draft manuscript is to provide an "existence proof" that it is possible to accomplish this.

Ok... but...

To use your example above,
|a> (|x1>+|x2> +...|xn>) + |b> (|y1> + ... ym>)
suppose the "binary measurement" is a spin measurement along the x-axis. We could suppose that the number of dimensions of the fine-grained measurement has something to do with the interaction between the particle and the SG apparatus. IOW, if the SG apparatus is oriented to measure along the x-axis, then the relevant "number of dimensions" is n and m (following your notation above). But if we rotate the SG apparatus so that it measures along some other axis, then the relevant number of dimensions becomes n' and m'.

The point is that we're not going to rotate the apparatus, but simply the initial state of the system to be measured. As such, the apparatus and environment and whatever that is going to do the measurement is IDENTICAL in the two cases. So if you have an argument of why we need n extra finegrained outcomes for |a> and m extra finegrained outcomes for |b> is to hold for the first case, it should also hold for the second case, because *the only thing that is changed is the to-be-measured state of the system, not the apparatus.
Whatever may be your reason to expect the n extra finegrained steps in the case we have |a> and the m extra finegrained steps in the case we have |b>, this entire measurement procedure will be resumed into A UNITARY INTERACTION OPERATOR that will split the relevant observer state into the n + m distinct states. A unitary interaction operator being a linear operator, it SHIFTS THROUGH the coefficients.

Let us take this again: let us assume that there are n+m distinct observer states that can result from the measurement of the "binary" measurement, namely the |x1> ...|yn> states (which now include the observer states which are to be distinct, and to which you can apply the APP). Of course, in his great naivity, the observer will lump together his n "x" states, and call it "a", and lump together his m "y" states, and call it "x" (post-coarse graining using the Kolmogorov additivity of probabilities).

But the evolution operator of the measurement apparatus + observer + environment and everything that could eventually matter (and that you will use for your argument of WHY there ought to be n |x_i> states and so on) is not supposed to DEPEND upon the incoming state. It is supposed to ACT upon the incoming state. If it were to DEPEND on it, and then ACT on it, it would be a non-linear operation ! Let us call it U.

So U(u1 |a> + u2|b>) results in the known state
|a> (|x1>+|x2> +...|xn>) + |b> (|y1> + ... ym>)

This means that U (|a>) needs to result in 1/u1 |a> (|x1>+|x2> +...|xn>)
and U (|b>) needs to result in 1/u2 |b> (|y1> + ... ym>)

(I took a shortcut here. In order to really prove it, one should first consider what U is supposed to do on |a> only, then on |b> only, and then on u1|a> + u2|b>, with the extra hypothesis that U(|a>) will not contain a component of |b> and vice versa - IOW that we have an ideal measurement)

This means that U(u2|a> + u1|b>) will result in what I said it would result in, namely:
u2/u1|a> (|x1>+|x2> +...|xn>) + u1/u2|b> (|y1> + ... ym>)

and this simply by the linearity of the U operator, which in turn is supposed to depend only on the measurement apparatus + environment + observer and NOT on the to-be-measured state. As this measurement environment is supposed to be indentical for both states, I don't see how you're going to wiggle out in this way (because from the moment you make it depend upon the to-be-measured state, you kill the unitarity (and even linearity) of the time evolution!)

The way to wiggle out is of course to say that the split does not occur in the measurement setup, but in the incoming state already! However, then you will meet another problem: namely decoherence. If you claim that the initial state is in fact, when we think that it is u1 |a> + u2|b>, in fact:
|a> (|x1>+|x2> +...|xn>) + |b> (|y1> + ... ym>)

then it is going to be difficult to show how we are going to obtain interference if ever we are now measureing |c> = |a> + |b> and |d> = |a> - |b>. Try it:
You'll find out that you will have to assume |x1> = |y1> etc... to avoid the in product being zero (decoherence!).

Now: (a|b) is supposed to be equal to u1 u2, or to: (sqrt(n)xsqrt(m))/(m+n)
But the in product of the two terms with the finegraining is going like m/(m+n) (if m is smaller than n). I don't see how you're going to get the right in products in all cases (all values of n and m) in this approach, unless I'm missing something.
 
  • #106
vanesch said:
Eh ? What progress ? That we can have any probability rule ? ... Now why can that extra rule not simply be: "use the Born rule" ?

Well that's all fine and good ... IF YOU'RE FEELING COMPLACENT :-p :biggrin:

Here's the idea: start with the APP, and justify it via an argument-by-symmetry (or envariance). Then, to recover the Born rule, add some additional rule or set of rules so that the Born rule and not some other rule emerges. Then, try to derive these newly postulated rules from GR. (Or if not GR, then something like it, maybe some alternate field theory.) iow, "work backwards" in baby steps from the APP, to some other set of rules, back to whatever the underlying theory is. Then clean up all the intermediate steps to make sure they're rigorous, etc etc. If not, start all over again.

Surely you can see that this would be a big payoff?:smile: :!)

Robin, Mike, and I have each (independently) proposed just such an alternate set of "extra rules" from which the Born rule is argued to emerge. Someday someone's going to hit paydirt!

I'll address the issue of unitarity in my next post.

David
 
  • #107
unitarity, linearity

Regarding the issue of unitarity: There are two possibilities: 1) the underlying theory is unitary; or, 2) the underlying theory is NOT unitary; and QM is unitary because it is an approximation of the underlying theory.

wrt my own scheme, I'm not entirely sure, although I'm leaning toward the latter case. Mike has argued that the underlying theory should be nonlinear. (which of course means not unitary.) In my own scheme, the fundamental entity is the state of the observer, which is specified in full via providing the metric g_{ij} and its power series at a point in spacetime. Since the power series has an infinite number of components, observer state space in my scheme is infinite dimensional (as might be expected/hoped). The operator that gives us the time-dependent evolution of the state of the observer is calculated by the parallel-transport law, ie by the geodesic equations:

<br /> \frac{d^{2}x^{k}}{dt^{2}} + <br /> \Gamma^{k}_{ij}<br /> \frac{dx^{i}}{dt}<br /> \frac{dx^{j}}{dt} = 0<br />

Assuming that spacetime is multiply connected, there will be more than one geodesic equation that is a solution to the above equations. Each separate geodesic is interpreted as an alternate possible evolution of the observer; ie, each separate geodesic exists in "superposition."

My point is that the above equation is nonlinear, so I'm thinking that my underlying scheme is fundamentally nonlinear, as Mike would argue it should be. The Born rule emerges in my scheme only if you make a particular approximation, in which a big chunk of the possible evolutions are approximated as being not there. (This is justified because this particular chunk has a very low probability, via the APP.)

But I'd like to understand your objection better -- it should help me understand why I should be so happy :blushing: that my scheme is nonlinear.

vanesch said:
The point is that we're not going to rotate the apparatus, but simply the initial state of the system to be measured. As such, the apparatus and environment and whatever that is going to do the measurement is IDENTICAL in the two cases.

I don't think I follow you here. If you are going to "rotate the initial state of the system to be measured," then there are several ways to do this, conceptually:

1) replace the initial system with a DIFFERENT system in a DIFFERENT (rotated) state; or

2) rotate the system (the particle), ie, PHYSICALLY; or

3) keep the particle in place, but rotate the SG apparatus, ie, by PHYSICALLY actually rotating the magnets.

However you do it, you are making a physical change of something: either the particle, or the apparatus. So you're not looking at the same overall (system+apparatus) state; you're looking at a DIFFERENT state. So why should you expect to go from one to the other via a unitary operator? I may just being dunce here, so keep hammering away ...

David
 
  • #108
straycat said:
Well that's all fine and good ... IF YOU'RE FEELING COMPLACENT :-p :biggrin:
Here's the idea: start with the APP, and justify it via an argument-by-symmetry (or envariance).

Just as a note: in the case of a symmetry argument, such as envariance, ALL probability rules have to give equal probabilities. This is, however, the ONLY case where we can deduce something about a hypothetical probability rule without postulating something about it.

Then, to recover the Born rule, add some additional rule or set of rules so that the Born rule and not some other rule emerges. Then, try to derive these newly postulated rules from GR. (Or if not GR, then something like it, maybe some alternate field theory.) iow, "work backwards" in baby steps from the APP, to some other set of rules, back to whatever the underlying theory is. Then clean up all the intermediate steps to make sure they're rigorous, etc etc. If not, start all over again.
Surely you can see that this would be a big payoff?:smile: :!)

I'm not sure it can even work in principle. Of course, for SOME situations, one can derive the Born rule in this (artificial) way, but I think that you cannot build it up as a general way ; as I tried to show with the in product example (although I just typed it like that ; I might have messed up, I agree that it is for the moment still "just intuition"). And if it succeeds, you need to postulate A LOT of unexplained physics!

Robin, Mike, and I have each (independently) proposed just such an alternate set of "extra rules" from which the Born rule is argued to emerge. Someday someone's going to hit paydirt!
I'll address the issue of unitarity in my next post.
David

I have all respect for the different attempts. As I think I proved, you do not need much as an extra hypothesis to derive the Born rule. I think that non-contextuality is a fair hypothesis ; I find the additivity of probabilities also a fair hypothesis (these are in fact very very close!). But they ALL need you to already say that SOME probability rule should emerge, and then you find WHICH ONE can emerge, satisfying the desired property.
So it is logically equivalent to say: "the probabilities that should emerge must follow the Born rule", and "the probabilities that should emerge must do so in a non-contextual (or additive) way". These are logically equivalent statements.

I find this fine. I don't see why one needs to go through this "equal probability stuff" (which is NOT non-contextual!) by postulating extra physics (unknown degrees of freedom) so that we miraculously obtain the Born rule, just to avoid to say that, well, the probability to be in a certain state is given by the Born rule, and not by equal probability. Why should the probabilities of the different worlds be equal for you to be in ?
 
  • #109
vanesch said:
And if it succeeds, you need to postulate A LOT of unexplained physics!
True ... but that is (hopefully) only a temporary state of affairs. The ultimate goal is that the "lot of unexplained physics" can be replaced by a very simple physical postulate -- like Einstein's equation, something like that.
vanesch said:
But they ALL need you to already say that SOME probability rule should emerge, and then you find WHICH ONE can emerge, satisfying the desired property.

So it is logically equivalent to say: "the probabilities that should emerge must follow the Born rule", and "the probabilities that should emerge must do so in a non-contextual (or additive) way". These are logically equivalent statements.

Again, the ultimate goal is that the assumption: "the probabilities that should emerge must follow the Born rule" is only a provisional one that is characteristic of a theory-in-progress. Once the "extra physics" mentioned above gets figured out, then the Born rule becomes derived, so need not be postulated.

vanesch said:
I find this fine. I don't see why one needs to go through this "equal probability stuff" (which is NOT non-contextual!) by postulating extra physics (unknown degrees of freedom) so that we miraculously obtain the Born rule, just to avoid to say that, well, the probability to be in a certain state is given by the Born rule, and not by equal probability. Why should the probabilities of the different worlds be equal for you to be in ?

Well you know the arguments at least as well as I do! We want the final finished product to avoid the charge of non-contextuality, to rely fundamentally on the APP. And we want to say that the Born rule, ie quantum mechanics, emerges from some underlying field theory, like GR -- and the ONLY probabililty assumption that we need to plug in is the APP, which is justified by a symmetry argument. The Born rule then gets "demoted," in a manner of thinking, because the APP is more fundamental to it.

I agree this is a difficult endeavor to undertake. But let's just suppose that it will work.How hard could it be? There are currently, what, hundreds of really smart mathematicians working on quantum gravity (string, loop, etc etc) -- and NONE of them afaik have incorporated the beautiful symmetry of the APP. Instead, they all ASSUME the (not symmetric, therefore ugly) Born rule to be the FUNDAMENTAL probability rule. (Smolin of course argues that string theory, in addition, makes the non-symmetric, therefore ugly, assumption of background dependence ... an argument to which I am sympathetic, although that is a whole differenet can of worms!) Perhaps the reason that all of these brilliant folks have not succeeded in figuring out quantum gravity is that their programmes are all tainted by one or both of the aforementioned assumptions. Surely the APP (like background independence) is a worthy symmetry principle into which some effort should be invested.

Don't forget that seemingly innocuous and simple symmetry principles have a long history of unexpectedly big payoffs. Einstein's principle of relativity being one such example!

David
 
  • #110
don't need no reeducation pleeez!

Tez said:
And in which context do you have a problem with it Howard? (...we are watching you... )

Hi Tez! Logic, me boy, logic... there's no logical implication that I have a problem with it in any context... really I don't, really... I didn't mean to imply I did... Please don't think I did! I don't need to be reeducated, really I don't!


Actually, I just meant to imply I'm not committed to probability *always* meaning subjective probability... in mathematics, it just refers to some things satisfying some formal axiomatic structure (s) ... so there, one is not committed to any view about its use in the world, or even whether it has any use... as far as probability in science, though, I'm a straight subjectivist a la Savage, even though I didn't bother to argue with the post about "unobservable preference orderings" vs. "measurable experimental outcomes"... yet.

Cheers,

H
 
  • #111
straycat said:
Surely the APP (like background independence) is a worthy symmetry principle into which some effort should be invested.

I fail to see what's so "symmetrical" about it:

You have the state a|u> + b|v>. If a and b are not equal, there is no symmetry between u and v, so I do not see why |u> should be equally probable to |v>. The "APP" arizes in those cases where a symmetry operator can swap the states ; but that's not the APP, it is *in the case of a symmetric state* that ALL probability rules need to assign equal probabilities to that particular case.
But I don't see why it is forbidden to say that the probability to observe |u> is not allowed to be a function of its hilbert coefficient ?
 
  • #112
vanesch said:
I fail to see what's so "symmetrical" about it:
You have the state a|u> + b|v>. If a and b are not equal, there is no symmetry between u and v, so I do not see why |u> should be equally probable to |v>. The "APP" arizes in those cases where a symmetry operator can swap the states ; but that's not the APP, it is *in the case of a symmetric state* that ALL probability rules need to assign equal probabilities to that particular case.

Well in the above, you have assumed standard QM, including the Born rule and the Hilbert space formalism. So of course you are correct that there is no symmetry. That's the point: the Born rule asserts asymmetry (except in special cases) wrt probabilities. The APP, otoh, is symmetrical wrt probabilities.

vanesch said:
But I don't see why it is forbidden to say that the probability to observe |u> is not allowed to be a function of its hilbert coefficient ?

I do not mean to imply that we are forced to accept the APP over the Born rule. ie, a symmetry argument is insufficient to forbid any non-symmetric alternative.

Take Einstein's principle of relativity. This is a symmetry principle, stating that the laws of physics are the same in all frames of reference. Does symmetry in and of itself mean we are "forbidden" to postulate frame-dependent laws? Well, no. (I could replace GR with Lorentzian relativity, for example.) All it means is that if I am trying to come up with new physics, and I have the choice between a frame-dependent and a frame-independent postulate, and I haven't yet worked out all the math so I don't know yet which one will work out, then as a betting man, I would put my money on the frame-independent one, all other considerations being equal (no pun intended :wink: ).

David
 
  • #113
vanesch said:
You have the state a|u> + b|v>.

There is another hidden assumption you are making here: that the number of "branches" associated with at measurement is in one-to-one correspondence with the different possible states of the observed system. In keeping with the spirit of Everett, the former should be equated with the number of physically distinct states into which the observer may evolve as a result of a measurement. The latter is equated with the number of physically distinct state of the observed system. In the standard treatment of the MWI, these two numbers are assumed to be the same. But why? It is conceivable that these may be different. So we have yet another independent postulate that is implicit to the standard MWI.

David
 
  • #114
straycat said:
Take Einstein's principle of relativity. This is a symmetry principle, stating that the laws of physics are the same in all frames of reference. Does symmetry in and of itself mean we are "forbidden" to postulate frame-dependent laws? Well, no.

Eh, yes ! That's exactly the content of the principle of relativity: we are forbidden to postulate frame-dependent laws!
 
  • #115
straycat said:
Well in the above, you have assumed standard QM, including the Born rule and the Hilbert space formalism. So of course you are correct that there is no symmetry. That's the point: the Born rule asserts asymmetry (except in special cases) wrt probabilities. The APP, otoh, is symmetrical wrt probabilities.

No, I didn't assume the Born rule. I just got out of unitary QM (on which we agree) that the state of the entire system (including myself) is a |u> + b|v>. One cannot say that the state is invariant under a swap of |u> and |v> which would be the statement of symmetry. In this case, there is no symmetry between the |u> and the |v> state, and I simply claimed that I don't see how a "symmetry" principle can assign now equal probabilities to |u> and to |v>. We can do so, by postulate, but it doesn't follow from any symmetry consideration. Given that doing so attracts us a lot of trouble, why do we do so ? And given that by assuming the Born rule we don't have that trouble, why not do so ?

I know of course where this desire comes from, and it is the plague of most MWI versions: one DOESN'T WANT TO SAY ANYTHING ABOUT PROBABILITIES. As such one seems to be locked up in the situation where we have a lot of independent terms (worlds) and we have to explain somehow how it comes that we only perceive ONE of them, given that we have "some state" in each of them. It seems indeed the most innocent to say that all these independent worlds are 'equally probable to be in', but that is nevertheless a statement about probability. Because without such a statement, we should be aware of ALL of our states, not just of one. So YOU CANNOT AVOID postulating somewhere a probability. My point is: pick the one that works ! Because there is no symmetry between those different worlds, we are NOT OBLIGED to say that they are equally probable. There is no physical symmetry between the low-hilbert norm states and the high hilbert norm states, in the sense that the state is not an eigenstate of any swap operator. But it is indeed tempting to state that all worlds are "equally probable" because that's how we've been raised into probabilities. We've been raised into counting the number of possible outcomes, and then counting the number of "successes", and taking the ratio. This is usually because we had genuinly symmetrical situations (like a dice) and in this case of course a symmetry argument implies that the probabilities are to be equal. So we do this also to the different terms in the wavefunction (although, I repeat, there is no a priori reason to take THIS distribution over another one, given that there is no symmetry in the situation). And so one has a *lot* of terms with small hilbert norm, and relatively few terms with high hilbert norm, and we can only find correspondence between the (observed) Born rule behaviour by "giving more weight" (1) to the few terms with high hilbert norm, or by "eliminating" (2) the vast lot of terms with small hilbert norm. Hence a lot of brain activity to find a mechanism to do so. You do (1) by trying to find extra physical degrees of freedom which we ignore, but split the few high-hilbert norm terms into a lot of small ones, so that their population is finally much larger than the initial small-hilbert norm states. Hanson does (2) by taking that worlds with very small hilbert norm are "mangled" (continue to suffer interactions with big terms), so that they somehow "don't count" in the bookkeeping.

And at the end of the day, you want to find the Born rule. Why not say so from the start ? :-p
 
  • #116
vanesch said:
Eh, yes ! That's exactly the content of the principle of relativity: we are forbidden to postulate frame-dependent laws!

I think you and I are merely speaking past each other a bit here. What I mean to say is that there is nothing that forbids us to say: "here is a symmetry principle, and despite its beauty, it's wrong." So, if we assume the principle of relativity is right, then we are forbidden from violating it; but we could, alternatively (and hypothetically), recognize the principle of relativity as being beautiful, but then turn around and not assume it. (Some people actually do this! ie proponents of Lorentzian relativity.)

My whole purpose in bringing up the principle of relativity is to compare it to the APP. Both are symmetry principles. With GR, of course, we assume the principle of relativity to be true. The question still remains, however, whether we should likewise adopt the APP as being true. I think that if the APP enables a succint derivation of quantum laws from deeper principles, in a manner that minimizes the total number of assumptions that must enter into the overall scheme, then the answer becomes "yes."

David
 
  • #117
vanesch said:
And at the end of the day, you want to find the Born rule. Why not say so from the start ? :-p

I think this whole discussion boils down to the fact that we see the APP differently. To you, the APP is no "better" than the Born rule, in the sense that we still need to postulate it. I can respect this PoV. However, there is a part of me that feels that the APP is sufficiently "natural" that it does not require an independent postulate.

I realize I sort of equivocate on this issue. Perhaps I should be more forceful in saying that the APP must be true. According to this view, the assumption of the Born rule in the standard formalism is a "band-aid" that we use because it "works," but which is ultimately unsatisfactory. Hence the need to replace it with the APP and some more physics. :rolleyes:

David
 
  • #118
vanesch said:
No ... I just got out of unitary QM (on which we agree) ...

Another instance of me equivocating ... I'm not sure I agree that the underlying theory must be characterized by unitary (= linear) evolution! See a few messages back.
 
  • #119
straycat said:
So, if we assume the principle of relativity is right, then we are forbidden from violating it; but we could, alternatively (and hypothetically), recognize the principle of relativity as being beautiful, but then turn around and not assume it.

Ah, ok! Yes, a wrong principle, no matter how beautiful, must not be adhered to :rolleyes:

Both are symmetry principles. With GR, of course, we assume the principle of relativity to be true. The question still remains, however, whether we should likewise adopt the APP as being true. I think that if the APP enables a succint derivation of quantum laws from deeper principles, in a manner that minimizes the total number of assumptions that must enter into the overall scheme, then the answer becomes "yes."

I agree with that, IF you have a (totally different!) theory, in which for some or other reason, there is a symmetry between the states, then you are right. But I fail to see how *the APP* is a "symmetry principle" in quantum theory. A symmetry principle should apply to the mathematical structure that is supposed to describe nature ; that is: it is an operator acting upon whatever set is supposed to be the set of possible states of nature (in QM, it is the rays of hilbert space ; in classical physics, it is the phase space, in GR, it is the set of all 4-dimensional manifolds that respect certain properties...), and that transforms it into the same state, or a state which has identical meaning (in the case of redundancy in the state space, like is the case with gauge symmetries).

I fail to see what operator can correspond to something which makes all states equivalent in a random state. The *envariance* symmetry of Zurek, on the other hand, IS a genuine symmetry (which is based upon the arbitrary phases in the tensor product of two subsystems).
Probabilities (just as any other observable phenomenon) that are to be derived from a state respecting a certain symmetry, should also obey the symmetry that is implemented. As such, I can understand that the probabilities, in the case of equal hilbert coefficients, must be equal (no matter what rule; a rule that does not obey it will run into troubles).

Of course, if we would now have some "APP quantum theory" in which we ALWAYS have, in each relevant basis, the same hilbert coefficients, then of course you are right. If you trim the hilbert space by a superselection rule that only allows for states in which the components are all equally long, so that all allowed-for states are "swappable", then of course ANY probability rule will be equivalent to the APP, and so will the Born rule. But I wonder how you're going to implement this ! This is very far from the original idea of quantum theory, and its superposition principle.
 
  • #120
vanesch said:
Of course, if we would now have some "APP quantum theory" in which we ALWAYS have, in each relevant basis, the same hilbert coefficients, then of course you are right. If you trim the hilbert space by a superselection rule that only allows for states in which the components are all equally long, so that all allowed-for states are "swappable", ...

To be honest I do not fully understand how Zurek can define "swappability" without letting some piece of QM -- and hence the Born rule! -- "sneak" in. iiuc, two states are swappable iff they have the same Hilbert space coefficients. Why couldn't we assume that they are swappable even if they have different coefficients? Because that would mean that they have different physical properties. So at the very least, Zurek is assuming that states must be elements of a Hilbert space, and that the Hilbert space coefficient is some sort of property characteristic of that state. Well if we are going to assume all that, we may as well just plug in Gleason's theorem, right? Or am I missing something?

vanesch said:
But I wonder how you're going to implement this !

Well you'll just have to read my paper! Which, btw, I am finally biting the bullet and submitting, today! :smile: :cool:

David
 
  • #121
straycat said:
To be honest I do not fully understand how Zurek can define "swappability" without letting some piece of QM -- and hence the Born rule! -- "sneak" in.

Well, Zurek accepts (of course) entirely the unitary part of QM, unaltered, and without "extra degrees of freedom". He introduces a unitary symmetry operator which "turns" randomly the phases of the basis vectors of system 1, and turns in opposite ways the phases of the basis vectors of system 2, and calls this an envariance symmetry operator. He argues that we can never measure the *phases* of the different basis vectors of system 1 (this comes from the redundancy in state description, namely the fact that a physical state corresponds to a RAY and not an element in hilbert space) or of system 2, and that, as such, his symmetry operator does not affect the physical state. He then goes on to enlarge the envariance symmetry operators, in which he swaps at the same time two states in the two hilbert spaces 1 and 2 (so that the overall effect of the Schmidt decomposition is simply to swap the hilbert coefficients), and notices that in the case of EQUAL COEFFICIENTS, this is a symmetry of the state.
He then introduces some assumptions (in that a unitary transformation of system 2 should not affect outcomes of system 1, including probabilities) and from some considerations arrives at showing that in such a case, all probabilities should be equal FOR THIS SPECIFIC STATE.

Why couldn't we assume that they are swappable even if they have different coefficients? Because that would mean that they have different physical properties. So at the very least, Zurek is assuming that states must be elements of a Hilbert space, and that the Hilbert space coefficient is some sort of property characteristic of that state.

Yes, in other words, he's accepting unitary quantum theory.

Well if we are going to assume all that, we may as well just plug in Gleason's theorem, right? Or am I missing something?

In order to derive Gleason's theorem, you have to make an extra assumption related to probabilities, which is the non-contextuality ; in other words, to assume that the probability of an outcome ONLY depends upon the properties of the component of the state within the compatible eigenspace corresponding to the desired outcome, and NOT on other properties of the state or the observable, such as THE NUMBER of different eigenspaces, and the components in the OTHER eigenspaces (of other, potential, outcomes). (the other possible outcomes, and their relation to the state, are the *context* of the measurement). As you know, the APP NEEDS this information: it needs to know HOW MANY OTHER EIGENSPACES have a non-zero component of the state in them. The Born rule doesn't: the length of the component in the relevant eigenspace is sufficient. And Gleason proves that the Born rule is THE ONLY rule which satisfies this property.

Zurek does something else, which is assuming additivity of the probabilities of the "fine-grained" state components and then uses the specific case where there are exactly a sufficient number of fine-grained state components to arrive at the requested Born rule. The request that, for this specific fine-grained situation, the probability of the component in the relevant (coarse-grained) eigenspace is given ONLY by the sum of "component probabilities" within this eigenspace, is yet another form of requiring non-contextuality: namely that the probability is entirely determined by the component in the relevant eigenspace (by taking the sum), and NOT by the context, which is how the rest is sliced up, and how the components are distributed over the rest. So in an involved way, he also requires non-contextuality. And then, by Gleason, you find the Born rule.
 
  • #122
vanesch said:
In order to derive Gleason's theorem, you have to make an extra assumption related to probabilities, which is the non-contextuality ; in other words, to assume that the probability of an outcome ONLY depends upon the properties of the component of the state within the compatible eigenspace corresponding to the desired outcome, and NOT on other properties of the state or the observable, such as THE NUMBER of different eigenspaces, and the components in the OTHER eigenspaces (of other, potential, outcomes). (the other possible outcomes, and their relation to the state, are the *context* of the measurement). As you know, the APP NEEDS this information: it needs to know HOW MANY OTHER EIGENSPACES have a non-zero component of the state in them. The Born rule doesn't: the length of the component in the relevant eigenspace is sufficient.

True, the length of the component is not the same thing as the number of other eigenspaces. However, the length is a normalized length, right? and the normalization process injects, does it not, information regarding the other eigenspaces? ie, instead of counting the total number of eigenspaces, we are adding up the "measure" of all the eigenspaces when we normalize.

David
 
  • #123
straycat said:
True, the length of the component is not the same thing as the number of other eigenspaces. However, the length is a normalized length, right? and the normalization process injects, does it not, information regarding the other eigenspaces? ie, instead of counting the total number of eigenspaces, we are adding up the "measure" of all the eigenspaces when we normalize.
David

Eh, yes. But the eigenspace, and its complement, are of course representing "the same" outcome (A or NOT A). So, ok, if you want to allow for non-normalized states, you need to allow for the eigenspace and its complement - I was assuming that we could normalize the states (and so does Gleason). We all know that the same physical state is represented by a ray in Hilbert space, so a common coefficient has no meaning and may just as well be normalized out. In fact, if the initial state is normalized, unitary time evolution will preserve this normalization. What counts is that the way that the complementary eigenspace is eventually sliced up or not, should not influence the outcome A to have non-contextuality.
 
  • #124
vanesch said:
Eh, yes. But the eigenspace, and its complement, are of course representing "the same" outcome (A or NOT A). So, ok, if you want to allow for non-normalized states, you need to allow for the eigenspace and its complement - I was assuming that we could normalize the states (and so does Gleason). We all know that the same physical state is represented by a ray in Hilbert space, so a common coefficient has no meaning and may just as well be normalized out. In fact, if the initial state is normalized, unitary time evolution will preserve this normalization.

Although in the case of successive measurements (say A -> B -> C), you have to renormalize with each measurement result. So if we want to assert that normalization happens only once, at the beginning, then we are restricting ourselves to a "no-collapse" framework. Let's suppose we want to know the conditional probability: what is the probability of outcome b', given outcome a' ? To answer this, we need to "collapse" onto outcome a', which means we have to recalculate the wavefunction, which includes a renormalization procedure. So ok, we could calculate the probability of b' using only unitary evolution (allowing A to remain superpositioned), but NOT if we want CONDITIONAL probabilities based on the outcome of A.
 
  • #125
straycat said:
Although in the case of successive measurements (say A -> B -> C), you have to renormalize with each measurement result. So if we want to assert that normalization happens only once, at the beginning, then we are restricting ourselves to a "no-collapse" framework. Let's suppose we want to know the conditional probability: what is the probability of outcome b', given outcome a' ? To answer this, we need to "collapse" onto outcome a', which means we have to recalculate the wavefunction, which includes a renormalization procedure. So ok, we could calculate the probability of b' using only unitary evolution (allowing A to remain superpositioned), but NOT if we want CONDITIONAL probabilities based on the outcome of A.

Bzzzzt ! The conditional probability P(a|b) is completely known if we know P(a and b) and P(b), because it is equal (by definition) to P(a and b)/P(b).

Now, imagine that the initial state is |psi0>, and that this first evolves into u1|b> + u2|c>, where of course |u1|^2 + |u2|^2 = 1 if psi0 was normalized. This means that we had probability |u1|^2 to observe |b> (so P(b) will be equal to |u1|^2).
Now, imagine that this further evolves into:
u1 (v1|a> + v2 |d>) |b> + u2 |c'>

Clearly, |v1|^2 + |v2|^2 = 1 too if the entire state is to stay normalized, which it will, through unitary evolution. If, after having observed |b>, we now observe |a>, we are in fact in the branch |a>|b>, which has hilbert norm u1 v1 and thus probability |u1|^2 |v1|^2. This is the probability to observe a and b, so P(a and b) = |u1|^2 |v1|^2

Applying our definition of conditional probability, we see that P(a|b) = |v1|^2, and we didn't have to re-normalize the state.
 
  • #126
vanesch said:
Bzzzzt ! The conditional probability P(a|b) is completely known if we know P(a and b) and P(b), because it is equal (by definition) to P(a and b)/P(b).

Now, imagine that the initial state is |psi0>, and that this first evolves into u1|b> + u2|c>, where of course |u1|^2 + |u2|^2 = 1 if psi0 was normalized. This means that we had probability |u1|^2 to observe |b> (so P(b) will be equal to |u1|^2).
Now, imagine that this further evolves into:
u1 (v1|a> + v2 |d>) |b> + u2 |c'>

Clearly, |v1|^2 + |v2|^2 = 1 too if the entire state is to stay normalized, which it will, through unitary evolution. If, after having observed |b>, we now observe |a>, we are in fact in the branch |a>|b>, which has hilbert norm u1 v1 and thus probability |u1|^2 |v1|^2. This is the probability to observe a and b, so P(a and b) = |u1|^2 |v1|^2

Applying our definition of conditional probability, we see that P(a|b) = |v1|^2, and we didn't have to re-normalize the state.


Aack! I think you are correct; my "conditional probability" critique was wrong. You caught me napping :zzz: .

There is still something I do not understand. Suppose we are measuring the spin state of a particle. If it is a spin 1/2 particle, then there are 2 states; spin 1, 3 states; etc. So when we apply the Schrodinger equation to the intial state |psi0>, it evolves into

u1|b> + u2|c> if the particle is spin 1/2, or

u1|b> + u2|c> + u3|e> if it is spin 1, etc.

So my question: how does the Schrodinger equation "know" how many states are possible? Is it part and parcel of our original definition of |psi0> ? Or does it somehow emerge from the Schrodinger equation itself, without our having to inject it externally?
 
  • #127
straycat said:
So my question: how does the Schrodinger equation "know" how many states are possible? Is it part and parcel of our original definition of |psi0> ?

?? I'd say it is part of the saying that it is a spin-1 particle in the first place. If it goes into 3 parts, we call it a spin-1 particle !

Sounds like: "how does a green car know it has to reflect green light ?" or something... unless I miss what you want to say.
 
  • #128
agreed

vanesch said:
BUT THAT IS NOTHING ELSE BUT NON-CONTEXTUALITY. It is always the same trick (equation 9a).cheers,
Patrick.


Yes, I agree completely. Deutsch, Wallace, Zurek etc do fine through the point of showing that equal-measure outcomes have equal probabilities. The next step involves assuming that probability is fixed after an experiment, and independent of observer. These are exactly the features which one does not find in APP. It's nice to show that Born can be derived from slightly weaker assumptions, but that doesn't mean that those assumptions follow from unitary QM. Worse, it doesn't mean that those assumptions are even consistent with unitary QM and our operational definition of probability.
 
  • #129
vanesch said:
Zurek does something else, which is assuming additivity of the probabilities of the "fine-grained" state components and then uses the specific case where there are exactly a sufficient number of fine-grained state components to arrive at the requested Born rule. The request that, for this specific fine-grained situation, the probability of the component in the relevant (coarse-grained) eigenspace is given ONLY by the sum of "component probabilities" within this eigenspace, is yet another form of requiring non-contextuality: namely that the probability is entirely determined by the component in the relevant eigenspace (by taking the sum), and NOT by the context, which is how the rest is sliced up, and how the components are distributed over the rest. So in an involved way, he also requires non-contextuality. And then, by Gleason, you find the Born rule.
Yeah- I thought you might be interested in these excerpts from a I comment wrote on the zurek argument. (I tried a bit to get it published there.) The point is that our arguments are almost identical, which is reassuring.


...Here I argue Zurek makes an implicit assumption which runs counter to the explicit assumptions.
...
The difficulty arises in extending the argument to decoherent outcomes whose measures are not equal, i.e. to density matrices whose non-zero diagonals are not equal. Here Zurek introduces a second environment, called C, and proposes that it is possible for E to become entangled with C in such a way that the density matrix for SC traced over E can (almost) be expressed by a collection of equal diagonal terms, with each diagonal term in the density matrix of S expanding into a finite set of equal diagonal terms in the density matrix of SC.. Now applying the swapping-symmetry argument to SC, Zurek gets that these SC outcomes must have equal probabilities.
Since the particular C-E entanglement required will occur on a set of measure zero of starting states, and cannot even approximately arise for the general case by any physical process represented by linear time evolution, the argument is not that such processes will occur but rather that they might sometimes occur, and the probabilities obtained in those special cases must be the same as those obtained in all cases because the density matrix for S is unaffected by the C-E* entanglement .
Treating the probabilities of S outcomes as sums over (more detailed) SC outcomes then gives the Born rule. This step, however, does not amount to simply using additivity of probabilities within a single probability space but rather implicitly assumes that the probabilities defined on S are simply related to the probabilities defined on SC. No matter how much that step accords with our experience-based common sense, it does not follow from the stated assumptions, which are deeply based on the idea that probabilities cannot be defined in general but only on a given system. Thus the question of why quantum probabilities take on Born values, or more generally of why they seem independent of where a line is drawn between system and environment, is not answered by Zurek's argument.
A counterexample may reinforce this point. A simple probability, defined from numbers* of diagonal terms in the density matrix of the system without weighting by measure, is entirely "envariant" and obeys the swapping symmetry. It does not obey any simple general relation between the probabilities defined on a system and on some supersystem. This probability is, of course, none other than 'world counting', which has frequently been argued to be the obvious probability to arise in collapse-free pictures without dynamical or metaphysical addenda. 2-8
Thus the problem of showing why the probabilities we experience emerge from quantum mechanics remains. ...
 
  • #130
vanesch said:
?? I'd say it is part of the saying that it is a spin-1 particle in the first place. If it goes into 3 parts, we call it a spin-1 particle !

Sounds like: "how does a green car know it has to reflect green light ?" or something... unless I miss what you want to say.

Sorry for the delayed reply ... been busy at work.

Let me see if I can explain my question. The operator for the (non-relativistic) Schrodinger equation in general form is

(1) H = T + V

If we are dealing with a spin 1/2 particle, then we know that the coefficients for the spin states take the form:

(2) |a_up|^2 = cos^2(theta), |a_down|^2 = sin^2(theta)

Obviously, these are normalized, meaning that sin^2 + cos^2 = 1 for any theta. But how exactly did we get from the general relation (1) to the specific relation (2)? Somewhere in this process, we had to inject the fact that there are *two* states. My point is that the Schrodinger equation does not tell us that there are two states; rather, this is an additional piece of information that is put in "externally" when we derive (2) from (1) so that we can normalize correctly. Therefore, the length of the component in the relevant eigenspace *does* depend on the total number of eigenspaces.

Unless I am missing something, which is entirely possible. (I have never understood the significance/meaning of "noncontextuality" -- hopefully I can fix that in this thread ...)

Just for reference, this is the statement that prompted my question:

vanesch said:
As you know, the APP NEEDS this information: it needs to know HOW MANY OTHER EIGENSPACES have a non-zero component of the state in them. The Born rule doesn't: the length of the component in the relevant eigenspace is sufficient.

David
 
  • #131
mbweissman said:
The difficulty arises in extending the argument to decoherent outcomes whose measures are not equal, i.e. to density matrices whose non-zero diagonals are not equal. ... This step, however ... rather implicitly assumes that the probabilities defined on S are simply related to the probabilities defined on SC. ...

I agree with this general line of reasoning. It seems so obvious that, when I first read Zurek's paper, it made me wonder whether I had missed some subtle point. So far I haven't found it though ...

David
 
  • #132
mbweissman said:
Treating the probabilities of S outcomes as sums over (more detailed) SC outcomes then gives the Born rule. This step, however, does not amount to simply using additivity of probabilities within a single probability space but rather implicitly assumes that the probabilities defined on S are simply related to the probabilities defined on SC.

EXACTLY !

No matter how much that step accords with our experience-based common sense, it does not follow from the stated assumptions, which are deeply based on the idea that probabilities cannot be defined in general but only on a given system. Thus the question of why quantum probabilities take on Born values, or more generally of why they seem independent of where a line is drawn between system and environment, is not answered by Zurek's argument.

Yes, that was also the reasoning I had. And IF you make that extra assumption (which, I think, corresponds to non-contextuality) then we *already know* that we will find the Born rule through Gleason's theorem.

A counterexample may reinforce this point. A simple probability, defined from numbers* of diagonal terms in the density matrix of the system without weighting by measure, is entirely "envariant" and obeys the swapping symmetry. It does not obey any simple general relation between the probabilities defined on a system and on some supersystem. This probability is, of course, none other than 'world counting', which has frequently been argued to be the obvious probability to arise in collapse-free pictures without dynamical or metaphysical addenda. 2-8
Thus the problem of showing why the probabilities we experience emerge from quantum mechanics remains. ...

Yes, your counter example is of course the "APP", which we should maybe give the name "RHA" (Revelator of Hidden Assumptions) :smile:
 
  • #133
straycat said:
Sorry for the delayed reply ... been busy at work.

Let me see if I can explain my question. The operator for the (non-relativistic) Schrodinger equation in general form is

(1) H = T + V

If we are dealing with a spin 1/2 particle, then we know that the coefficients for the spin states take the form:

(2) |a_up|^2 = cos^2(theta), |a_down|^2 = sin^2(theta)

Obviously, these are normalized, meaning that sin^2 + cos^2 = 1 for any theta. But how exactly did we get from the general relation (1) to the specific relation (2)? Somewhere in this process, we had to inject the fact that there are *two* states.

We don't derive (2) from (1). (2) is part of the interpretation of the Hilbert space of states. In order to set up a quantum theory, we need to start NAMING the independent degrees of freedom, and *assign* them operational meanings. This is a part that is often skipped in textbooks, because they are usually interested in quantizing classical systems, and in this case, the independent degrees of freedom are given by the points in configuration space of the classical system, so it is "automatic". And then we STILL need to put in, by hand, a few extra degrees of freedom, such as spin.
Things like (2) are usually called the "kinematics" of the theory, while (1) is the "dynamics".

This compares, in classical physics, to: (1) specifying the number of particles and degrees of freedom of the mechanical system and (2), writing down Newton's equation or its equivalent.
Clearly, you cannot derive the number of planets from Newton's equation, and you cannot derive the number of degrees of freedom for spin from the Schroedinger equation.
 
  • #134
vanesch said:
...
This compares, in classical physics, to: (1) specifying the number of particles and degrees of freedom of the mechanical system and (2), writing down Newton's equation or its equivalent.

Clearly, you cannot derive the number of planets from Newton's equation, and you cannot derive the number of degrees of freedom for spin from the Schroedinger equation.

This is a good analogy. To calculate the orbital of a planet (say Mars), we take as input the masses, intial positions, and initial velocities of the sun, planets, and other objects in the solar system, and plug these into Newton's equation, which spits out the solution x_mars(t). (We could then talk about the "Isaac rule" which states that "x(t) is interpreted as the trajectory.") Likewise, to calculate the probability associated with a given observation, we take the initial state of the system, which includes stuff like the particle's spin, which includes the total number of possible spin states, and plug this into the Schrodinger equation, which spits out the amplitudes a_n. Finally, we apply the Born rule, which tells us that |a|^2 gives us the probability.

Now you said earlier that

vanesch said:
As you know, the APP NEEDS ... to know HOW MANY OTHER EIGENSPACES have a non-zero component of the state in them. The Born rule doesn't: the length of the component in the relevant eigenspace is sufficient.

So you are saying that the probability of, say, spin up is independent of the total number of eigenstates (two in this case), because it depends only upon a_up. But isn't this like saying that the trajectory of Mars is independent of the trajectories of the other planets, because it depends only on x_mars(t)? Obviously this statement is false, because you cannot change (say) x_jupiter(t) without inducing a change in x_mars(t) as well.

To sum up: if we assume non-contextuality, this means (correct me if I am wrong) that we are assuming that the probability of the n^th outcome depends only on a_n and not on anything else, such as the total number N of eigenspaces. My difficulty is that I do not see how this assumption is tenable, given that you cannot calculate a_n without knowing (among other things) the total number of eigenspaces. It would be analogous to the assumption that the trajectory of Mars is independent of the trajectory of Jupiter.

So what am I missing?

David
 
  • #135
straycat said:
This is a good analogy. To calculate the orbital of a planet (say Mars), we take as input the masses, intial positions, and initial velocities of the sun, planets, and other objects in the solar system, and plug these into Newton's equation, which spits out the solution x_mars(t). (We could then talk about the "Isaac rule" which states that "x(t) is interpreted as the trajectory.")

Worse! The number of planets will even CHANGE THE FORM of Newton's equation: the number of 1/r^2 terms in the force law for each planet will be different to whether there are 2, 3, 5, 12, 2356 planets. In other words, the DIMENSION OF THE PHASE SPACE defines (partly) the form of Newton's equation.

Likewise, to calculate the probability associated with a given observation, we take the initial state of the system, which includes stuff like the particle's spin, which includes the total number of possible spin states, and plug this into the Schrodinger equation, which spits out the amplitudes a_n.

In the same way, the FORM of the Schroedinger equation (or better, of the Hamiltonian) will depend upon the spins of the particles ; and this time not only the number of terms, but even the STRUCTURE of the Hamiltonian: a Hamiltonian for a spin-1/2 particle CANNOT ACT upon the Hilbert space of a spin-1 particle: it is not the right operator on the right space.

So you are saying that the probability of, say, spin up is independent of the total number of eigenstates (two in this case), because it depends only upon a_up. But isn't this like saying that the trajectory of Mars is independent of the trajectories of the other planets, because it depends only on x_mars(t)? Obviously this statement is false, because you cannot change (say) x_jupiter(t) without inducing a change in x_mars(t) as well.

I was referring to the number of eigenspaces of the hermitean measurement operator, NOT about the unitary dynamics (which is of course sensitive to the number of dimensions of HILBERTSPACE, which is nothing else but the number of physical degrees of freedom). The eigenspaces of the hermitean measurement operator, however, depend entirely on the measurement to be performed (the different, distinguishable, results). When the measurement is complete, then all eigenspaces are one-dimensional. But clearly that's not possible, because that means an almost infinite amount of information.
The hermitean measurement operator is in fact just a mathematically convenient TRICK to say which different eigenspaces of states correspond to distinguishable measurement results (and to include a name = real number for these results). What actually counts is the slicing up of the hilbert space in slices of subspaces which "give the same results".

The extraction of probabilities in quantum theory, comes from the confrontation of two quantities:
1) the wavefunction |psi> and 2) the measurement operator (or better, the entire set of eigenspaces) {E_1,E_2...E_n},
and the result of this confrontation has to result in assigning a probability to each distinct outcome.

I tried to argue in my paper that every real measurement always has only a finite number of eigenspaces {E_1...E_n} ; that is to say, the result of a measurement can always be stored in a von Neumann computer with large, but finite, memory. You can try to fight this, and I'm sure you'll soon run into thermodynamical problems (and you'll even turn into a black hole :biggrin: ).

As such, when I do a specific measurement, so when I have a doublet: {|psi>,{E_1...E_n}}, then I need to calculate n numbers, p_1,... p_n, which are the predicted probabilities of outcomes associated, respectively, to E_1 ... E_n.

This means that p_i ({|psi>,{E_1...E_n}}) in general.
This means that p_i can change completely if we change ANY of the E_j. On the other hand, FOR A GIVEN SET OF {E_1...E_n}, they have to span a Kolmogorov probability measure. But FOR A DIFFERENT SET, we can have ANOTHER probability measure.

The Born rule says that p_i is given by <psi|P_i|psi> (if psi is normalized), where P_i is the projector on E_i. The APP says that p_1 = 1/k, with k the number of projectors which do not annihilate |psi>.

Non-contextuality is the claim that p_i can only depend upon |psi> and E_i. Gleason's theorem says then, that IF WE REQUIRE that p_i is ONLY a function of |psi> and E_i (no matter what the other E_k are, and how many there are), then the ONLY solution is the Born rule. If the probability for a measurement to say that the state is in E_i can only depend upon E_i itself, and the quantum state |psi>, then the only solution is the Born rule.

This rules out the APP, for the APP needs to know the number k of eigenspaces which contain a component of |psi>. The APP is hence a non-non-contextual rule. It needs to know 'the context' of the measurement in WHICH we are trying to calculate the probability of |psi> to be in E_i.

To sum up: if we assume non-contextuality, this means (correct me if I am wrong) that we are assuming that the probability of the n^th outcome depends only on a_n and not on anything else, such as the total number N of eigenspaces. My difficulty is that I do not see how this assumption is tenable, given that you cannot calculate a_n without knowing (among other things) the total number of eigenspaces. It would be analogous to the assumption that the trajectory of Mars is independent of the trajectory of Jupiter.

No, that's because you're confusing two different sets of dimensions. The physical situation determines the number of dimensions in Hilbert space, and the dynamics (the unitary evolution) is dependent upon that only. But there's no discussion about the number of dimensions of Hilbert space (the number of physical degrees of freedom).
The number of EIGENSPACES is related to the resolution and kind of the measurement, in that many physical degrees of freedom will give identical measurement results, which are then lumped into ONE eigenspace. It is more about HOW WE DISTRIBUTE the physical degrees of freedom over the DIFFERENT measurement outcomes, and how this relates to the probability of outcome.
Non-contextuality says that if we perform TWO DIFFERENT measurements, but which happen to have a potential outcome in common (thus, have one of their E_i in common), that we should find the same probability for that outcome, for the two different cases. It is a very reasonable requirement at first sight.
 
  • #136
Sorry again for the much delayed response.

I've been pondering your last post. It seems you are drawing a distinction between (1) the number of dimensions of Hilbert space and (2) the number of eigenspaces of the measurement operator. I realize these are different, but there is still something I am not quite grokking. I'll have to re-read this thread and ponder more when I get more time. In the meantime, a few comments/questions.

vanesch said:
Non-contextuality is the claim that p_i can only depend upon |psi> and E_i. ...

... the APP needs to know the number k of eigenspaces which contain a component of |psi>. The APP is hence a non-non-contextual rule.

Suppose we assume the APP. Given a particular measurement to be performed, suppose we have K total fine-grained outcomes, with k_i the number of fine-grained outcomes corresponding to the i^th coarse-grained result. eg, we have N position detector elements, i an integer in [1,N], and the sum of k_i over all i equals K. So the probability of detection at the i^th detector element is k_i / K, and we define:
E_i = k_i / K

So if I claim that p_i can depend only upon E_i (ie p_i = E_i), then it seems to me that I could argue, using the same reasoning that you use above for the Born rule, that the APP is non-contextual. What is wrong with my reasoning? I suppose you might say that you cannot calculate E_i (in the framework of the APP) without knowledge of K, ie without knowledge of the context. But it still seems to me that you likewise cannot calculate E_i in the framework of the Born rule, without knowledge of the measurement operator. iow, I'm trying to argue that the APP and the Born rule are either both contextual, or both non-contextual, depending on how exactly you define contextual, and you can't distinguish them based on "contextuality."

Perhaps I should study Gleason's theorem in greater detail than I have done so far. I actually think it is somewhat remarkable that it leads to the Born rule. However, it still seems to me that the assumption of a normalized Hilbert space for state representation is where the Born rule sneaks in. That is, Hilbert space assumes the state is represented by f, and the sum of |f|^2 over all states equals 1 (by normalization). So really it's not that surprising that |f|^2 is the only way to get a probability.

vanesch said:
It is a very reasonable requirement at first sight.

Do you say "at first sight" because a careful analysis indicates that it's not all that reasonable?

vanesch said:
I tried to argue in my paper that every real measurement always has only a finite number of eigenspaces {E_1...E_n} ; that is to say, the result of a measurement can always be stored in a von Neumann computer with large, but finite, memory. You can try to fight this, and I'm sure you'll soon run into thermodynamical problems (and you'll even turn into a black hole :biggrin: ).

I actually agree with you here. The argument in my mind goes like this: consider a position measurement. If you want to come up with a continuous measurement variable, this would be it. But from a practical perspective, a position measurement is performed via an array or series of discrete measurement detectors. The continuous position measurement is then conceived as the theoretical limit as the number of detector elements becomes infinite. But from a practical, and perhaps from a theoretical, perspective, this limit cannot ever be achieved: the smallest detector element I can think of would be (say) an individual atom, for example the atoms that make up x-ray film.

David
 
  • #137
straycat said:
I've been pondering your last post. It seems you are drawing a distinction between (1) the number of dimensions of Hilbert space and (2) the number of eigenspaces of the measurement operator. I realize these are different,

Yes, this is essential. The number of dimensions in Hilbert space is given by the physics, and by physics alone, of the system, and might be very well infinite-dimensional. I think making assumptions on the finiteness of this dimensionality is dangerous. After all, you do not know what degrees of freedom are hidden deep down there. So we should be somehow independent of the number of dimensions of the Hilbert space.

However, the number of eigenspaces of the measurement operator is purely determined by the measurement apparatus. It is given by the resolution by which we could, in principle, determine the quantity we're trying to measure, using the apparatus in question. You and I agree that this must be a finite number, and a rather well-determined one. This is probably where we are differing in opinion, and where you seem to claim "micromeasurements" of eventually unknown physics of which we are not aware versus "macromeasurements" which are just our own coarse-graining of these micromeasurements- while I claim that with every specific measurement goes a certain, well-defined number of outcomes (which could eventually be more fine-grained than the observed result but that this should not be dependent on "unknown physics", but that a detailled analysis of the measurement setup should reveil that to us). I would even claim that a good measurement apparatus makes the observed number of outcomes about equal to the real number of eigenspaces.

Suppose we assume the APP. Given a particular measurement to be performed, suppose we have K total fine-grained outcomes, with k_i the number of fine-grained outcomes corresponding to the i^th coarse-grained result. eg, we have N position detector elements, i an integer in [1,N], and the sum of k_i over all i equals K. So the probability of detection at the i^th detector element is k_i / K, and we define:
E_i = k_i / K

Yes, but I'm claiming now that for a good measurement system, k_i = 1 for all i, and even if it isn't (for instance, you measure with a precision of 1 mm, and your numerical display only displays up to 1 cm resolution), you're not free to fiddle with k_i as you like.
Also, you now have a strange outcome! You ALWAYS find probability E_i for outcome i, no matter what was the quantum state ! Even if the quantum state is entirely within the E_i eigenspace, you'd still have a fractional probability ? That would violate the rule that two measurements applied one after the other will give the same result.

So if I claim that p_i can depend only upon E_i (ie p_i = E_i), then it seems to me that I could argue, using the same reasoning that you use above for the Born rule, that the APP is non-contextual.

No, non-contextuality has nothing to do with the number E_i you're positioning here, it is a property of being only a function of the eigenspace (spanned by the k_i subspaces) and the quantum state, no matter how the other eigenspaces are sliced up. Of course, in a way, you're right: if the outcome is INDEPENDENT on the quantum state (as it is in your example), you are indeed performing a non-contextual measurement. In fact, the outcome has nothing to do with the system: outcome i ALWAYS appears with probability E_i. But I imagine that you only want to consider THOSE OUTCOMES i THAT HAVE A PART OF the quantum state in them, right ? And THEN you become dependent on what happens in the other eigenspaces.

That is, Hilbert space assumes the state is represented by f, and the sum of |f|^2 over all states equals 1 (by normalization). So really it's not that surprising that |f|^2 is the only way to get a probability.

Well, there's a difference in the following sense: if you start out with a normalized state, you will always keep a normalized state under unitary evolution, and if you change basis (change measurement), you can keep the same normalized vector. That cannot be said for the E_i and k_i construction, which needs to be redone after each evolution, and after each different measurement basis.

Do you say "at first sight" because a careful analysis indicates that it's not all that reasonable?

Well, it is reasonable, but it is an EXTRA assumption (and, according to Gleason, logically equivalent to postulating the Born rule). It is hence "just as" reasonable as postulating the Born rule.
What I meant with "at first sight" is that one doesn't realize the magnitude of the step taken! In unitary QM, there IS no notion of probability. There is just a state vector, evolving deterministically by a given differential equation of first order, in a hilbert space. From the moment that you require, no matter how little, a certain quality of a probability issued from that vector, you are in fact implicitly postulating an entire construction: namely that probabilities ARE going to be generated from this state vector (probabilities for what, for whom?), that only part of the state vector is going to be observed (by whom?) etc... So the mere statement of a simple property of the probabilities postulates in fact an entire machinery - which is not obvious at first sight. Now if your aim is to DEDUCE the appearance of probabilities from the unitary machinery, then implicitly postulating this machinery is NOT reasonable, because it implies that you are postulating what you were trying to deduce in one way or another.

I actually agree with you here. The argument in my mind goes like this: consider a position measurement. If you want to come up with a continuous measurement variable, this would be it. But from a practical perspective, a position measurement is performed via an array or series of discrete measurement detectors. The continuous position measurement is then conceived as the theoretical limit as the number of detector elements becomes infinite. But from a practical, and perhaps from a theoretical, perspective, this limit cannot ever be achieved: the smallest detector element I can think of would be (say) an individual atom, for example the atoms that make up x-ray film.

That's what I meant, too. There's a natural "resolution" to each measurement device, which is given by the physics of the apparatus. An x-ray film will NOT be in different quantum states for positions which differ much less than the size of an atom (or even a bromide xtal). This is not "unknown physics" with extra degrees of freedom. I wonder whether a CCD type camera will be sensing on a better resolution than one pixel (meaning that the quantum states would be different for hits at different positions on the same pixel). Of course, there may be - and probably there will be - some data reduction up to the display, but one cannot invent, at will, more fine-grained measurements than the apparatus is actually naturally performing. And this is what determines the slicing-up of the Hilbert space in a finite number of eigenspaces, which will each result in macroscopically potentially distinguishable "pointer states". And I think it is difficult (if not hopeless) to posit that these "micromeasurements" will arrange themselves each time in such a way that they work according to the APP, but give rise to the Born rule on the coarse-grained level. Mainly because the relationship between finegrained and coarse grained is given by the measurement apparatus itself, and not by the quantum system under study (your E_i = k_i/K is fixed by the physics of the apparatus, independent of the state you care to send onto it ; the number of atoms on the x-ray film per identified "pixel" on the scanner is fixed, and not depending on how it was irradiated).

cheers,
Patrick.
 
  • #138
You can try to fight this, and I'm sure you'll soon run into thermodynamical problems (and you'll even turn into a black hole :biggrin:).
Proof by threat of black hole!
 
  • #139
Hurkyl said:
Proof by threat of black hole!

I'm proud to have found a new rethorical technique :biggrin:
 
  • #140
It takes a singular mind to come up with such things! (Okay, I'll stop now)
 
  • #141
mbweissman said:
Treating the probabilities of S outcomes as sums over (more detailed) SC outcomes then gives the Born rule. This step, however, does not amount to simply using additivity of probabilities within a single probability space but rather implicitly assumes that the probabilities defined on S are simply related to the probabilities defined on SC. No matter how much that step accords with our experience-based common sense, it does not follow from the stated assumptions, which are deeply based on the idea that probabilities cannot be defined in general but only on a given system. Thus the question of why quantum probabilities take on Born values, or more generally of why they seem independent of where a line is drawn between system and environment, is not answered by Zurek's argument.

Hi Michael,

5 or so years ago when I was visiting Paul Kwiat you gave me a preprint of how you thought the Born rule could/should be derived. I remember there was a cute idea in there somewhere, though I can't remember what it was! How did it pan out?

Tez
 
  • #142
vanesch said:
And I think it is difficult (if not hopeless) to posit that these "micromeasurements" will arrange themselves each time in such a way that they work according to the APP, but give rise to the Born rule on the coarse-grained level. Mainly because the relationship between finegrained and coarse grained is given by the measurement apparatus itself, and not by the quantum system under study (your E_i = k_i/K is fixed by the physics of the apparatus, independent of the state you care to send onto it ; the number of atoms on the x-ray film per identified "pixel" on the scanner is fixed, and not depending on how it was irradiated).

Well, maybe it's not as difficult / hopeless as you might think! Let's play around for a moment with the idea that all measurements boil down to one particle interacting with another. That is, the fundamental limit of resolution of a particle detector is governed by the fact that the detector is made of individual particles. So if we look at the micro-organization at the fine-grained level, we see micro-structure that is determined by the properties of the particles in question; let's say, some property that is characteristic of fermions / bosons for fermi /bose statistics, respectively. When a particle hits an atom in a CCD detector, then there is a corresponding micro-structure that always follows some particular pattern, and it gives rise to the Born rule when you look at it from a coarse-grained perspective. So if particles were "constructed" differently, then we might not have the Born rule, we might have some other rule. This, in fact, is exactly how my toy scheme works!

This view is consistent with the notion that it does not matter whether there is data reduction up to the display. That is, it does not matter whether the CCD has resolution of 1 mm or 1 cm; if two different CCD's have different pixel resolution, but are made of the same types of atoms, then they will have the same fundamental fine-grained "resolution" when we look at the micro-structure.

I'm starting to contemplate a thought experiment, not sure where it will take me. Suppose we have a CCD camera (length, say, 10 cm) and we remove a 2 cm chunk of it which we replace with a lens that focuses all particles that would have hit the plate on that 2 cm stretch onto (for the sake of argument) a single atom. What effect do we expect this will have on our measurement probabilities? Contrast that to a different scenario: we have a CCD camera, length 10 cm, with resolution 1 mm. Remove a 2 cm chunk and replace it with a single pixel, ie 2 cm resolution. But both CCD setups are made of the same types of atoms. I would expect that the probability of detection over the 2 cm single pixel equals the sum of the probability of detection of all 20 of the individual 1 mm pixels; my reasoning is that in both setups, we have the same density and type of atoms in the CCD's. But I would imagine that using the lens setup, we would get something completely different, since we are effectively replacing detection over a 2 cm stretch using lots of atoms with detection using only one atom.
 
  • #143
straycat said:
I'm starting to contemplate a thought experiment, not sure where it will take me. Suppose we have a CCD camera (length, say, 10 cm) and we remove a 2 cm chunk of it which we replace with a lens that focuses all particles that would have hit the plate on that 2 cm stretch onto (for the sake of argument) a single atom. What effect do we expect this will have on our measurement probabilities? Contrast that to a different scenario: we have a CCD camera, length 10 cm, with resolution 1 mm. Remove a 2 cm chunk and replace it with a single pixel, ie 2 cm resolution. But both CCD setups are made of the same types of atoms. I would expect that the probability of detection over the 2 cm single pixel equals the sum of the probability of detection of all 20 of the individual 1 mm pixels; my reasoning is that in both setups, we have the same density and type of atoms in the CCD's. But I would imagine that using the lens setup, we would get something completely different, since we are effectively replacing detection over a 2 cm stretch using lots of atoms with detection using only one atom.

Actually this reminds me of the quantum zeno effect ( http://en.wikipedia.org/wiki/Quantum_Zeno_effect ), which I mentioned in post #37 of this thread. From the wiki description, the experiment they do is sort of similar to the thought experiment I outlined above, except that I am playing around with resolution of the position measurement, whereas they were playing around with the resolution of a time measurement in the experiment described in wiki. The point of the zeno effect is that if you change the resolution of the time measurement at the fine grained level, then you change the probability distribution as a function of time. Similarly, I expect that if you change the resolution of the position measurement in a fundamental sense, ie using the lens setup, then you should change the probability distribution as a function of position. But if you simply swap the 1 mm pixel with the 2 cm pixel, then (I expect) you will not change the probability as a function of position, because you have done nothing to change the fundamental micro-structure, since the 1 mm and 2 cm CCD detectors have the same density of atoms.
 
  • #144
Doesn't your spatial-resolution fiddling bear a family resemblance to Asfhar's analysis? I believe that was described here recently in Quantum Zeno terms.
 
  • #145
selfAdjoint said:
Doesn't your spatial-resolution fiddling bear a family resemblance to Asfhar's analysis? I believe that was described here recently in Quantum Zeno terms.

Hmm, I've never thought of comparing the two. It's been a long time since I've thought about the Afshar experiment. I always belonged to the camp that thought there was a flaw somewhere in his analysis, though. That is, I tend to think that the CI and the Everett interpretation each make exactly the same predictions as any of the other formulations of QM (see, eg, the wonderful paper [1]) -- so I am a bit biased against Afshar's (and Cramer's) claims to the contrary.

As for the Zeno effect, I have actually not really pondered it really really deeply. But from my cursory contemplation, the existence of the Zeno effect does not surprise me all that much. To me, the main lesson of the Zeno effect could be stated loosely: how you measure something (the resolution of the time measurements) has an effect on the probability distribution (probability of decay as a function of time). But that is simply Lesson # 1 (in my mind) in quantum mechanics. eg, the 2-slit exp tells us that how we measure something (whether we do or do not look at the slits) has an effect on the resulting probability distribution (where it hits the screen). So perhaps the Zeno effect is just teaching us the same lesson as the 2-slit exp, but dressed up differently.

So my knee jerk reaction to your question would be that Afshar's analysis is based in a (somehow) flawed reading/implementation of the CI (and MWI), but the Zeno effect is founded upon a correct implementation of quantum theory. I'd have to take another look at Afshar though to see the comparison with Zeno ...

David

[1] Styer et al. Nine formulations of quantum mechanics. Am J Phys 70:288-297, 2002
 
  • #146
selfAdjoint said:
Doesn't your spatial-resolution fiddling bear a family resemblance to Asfhar's analysis? I believe that was described here recently in Quantum Zeno terms.

Oh yea! Afshar used a lens in his setup too -- now I remember -- duhhh :blushing:

D

http://en.wikipedia.org/wiki/Afshar_experiment
 
  • #147
wow I just the wiki article on the ashfar experiment... mmm.. so proc. spie is an optical engineering journal and not a physics journal...

I guess it must be generally believed by the physics powers that be that ashfar's interpretation of the experiment is erroneous.

good enough for me i guess.. hehe
 
  • #148
alfredblase said:
I guess it must be generally believed by the physics powers that be that ashfar's interpretation of the experiment is erroneous.

Yea, I just linked from wiki to Lubos Motl's blog article [1] criticising the Afshar analysis, and I see with some amusement that Lubos' critique is essentially the same critique that I made myself [2] over in undernetphysics ... except that I made my critique a month earlier!

That's right, I beat him to the punch ... who's ya' daddy now? :cool:

DS <ducking in case Lubos is lurking about somewhere ...>

[1] http://motls.blogspot.com/2004/11/violation-of-complementarity.html

[2] http://groups.yahoo.com/group/undernetphysics/message/1231
 
  • #150
non-linear decoherence

Tez said:
Hi Michael,

5 or so years ago when I was visiting Paul Kwiat you gave me a preprint of how you thought the Born rule could/should be derived. I remember there was a cute idea in there somewhere, though I can't remember what it was! How did it pan out?

Tez


Hi Tez- Sorry for the delay- haven't been checking the forum. The idea was that if there were a non-linear decoherence process, the proper ratio of world-counts could arise asymptotically without fine-tuning. Basically it runs like this: if large-measure branches decohere faster than small-measure ones the limiting steady-state distributions would have the same average measure per branch. Hence branch count is simply proportional to measure.

How'd it work out? It was published in Found Phys. Lett., after some extraordinary constructive criticism from a referee. So far there are no obvious holes in it- e.g. no problem with superluminal communication, unlike some types of non-linear dynamics. On the other hand, it proposes extra machinery not in ordinary quantum mechanics, without giving a specific theory. Although the extra gunk is much less Rube-Goldbergish than in explicity collapse theories, it would be nice not to have to propose something like that at all.

I'm about to post a follow-on, in which I point out that once the non-linear processes have been proposed to rescue quantum measurement, they give the second law at no extra cost. A similar point was made by Albert in his book Time and Chance, but he was referring to non-linear collapse (much uglier) rather than simple non-linear decoherence.
 
Back
Top