PeterDonis said:
(b) Thermal interpretation: The q-expectation of the measurement is zero (an equal average of +1 and -1), but each individual measurement gives an inaccurate result because of the way the measurement/detector are constructed, so only the average over many results on an ensemble of identically prepared electrons will show the q-expectation. For each individual measurement, random nonlinear fluctuations inside the detector cause the result to be either +1 or -1.
I might be misunderstanding you, but I actually think (b) is not what
@A. Neumaier is saying, at least if by "random nonlinear fluctuations" you mean that there's some change to the unitary dynamics underlying quantum mechanics. Rather, he's saying that the nonlinearity comes from coarse-graining, that is, from neglecting some details of the state, which would actually evolve linearly if you could somehow add those details back in.
This was my reading from the beginning, and is the source of my question. I feel quite stupid right now and that I must be missing something obvious, but I'm going to press on anyway and try to be more specific about my confusion.
----------------------------------
Let's start with the setup in Section 3 of
the fourth TI paper, where we've written the Hilbert space of the universe as ##H=H^S\otimes H^E## where ##H^S## is two-dimensional, and assume our initial state is ##\rho_0=\rho^S\otimes\rho^E##. We have some observable ##X^E## on ##H^E##, and we're thinking of this as being something like the position of the tip of a detector needle. Using thermal interpretation language, we can say that we're interested in the "q-probability distribution" of ##X^E## after running time forward from this initial state; this can be defined entirely using q-expectation values, so I think
@A. Neumaier and I agree that this is a physically meaningful object to discuss. If, after some experiment, the q-probability distribution of ##X^E## has most of its mass near some particular ##x\in\mathbb{R}##, then I'm happy to say that there's no "measurement problem" with respect to that experiment.
Consider two state vectors ##\psi_1## and ##\psi_2##, and suppose they are orthogonal, and pick some initial density matrix ##\rho^E## for the environment. Suppose that:
(i) starting in the state ##\psi_1\psi_1^*\otimes\rho^E## and running time forward a while yields a q-probability distribution for ##X^E## with a single spike around some ##x\gg 0##, and
(ii) similarly with ##\psi_2## around ##-x##.
The question then is: what does the q-probability distribution of ##X^E## look like if we instead start with ##\frac12(\psi_1+\psi_2)(\psi_1+\psi_2)^*\otimes\rho^E##?
The two competing answers I'm worried about are:
(a) It will be bimodal, with a peak around ##x## and a peak around ##-x##
(b) It will be unimodal, concentrated around ##-x## or around ##x##, with the choice between the two depending in some incredibly complicated way on the exact form of ##\rho_E##. (In this story, maybe there's a choice of ##\rho^E## that will give something like (a), but it would require a ludicrous amount of symmetry and so there's no need to worry about it.)
The reason I'm confused, then, is that I thought that the decoherence story involves (among other things)
deducing (a) from (i) and (ii). In particular, I thought it followed from the linearity of time evolution together with the whole business with decaying off-diagonal terms in the density matrix, but I don't understand the literature enough to be confident here.
Am I just wrong about what the decoherence story claims? Is it just that they assume enough symmetry in ##\rho^E## to get (a) to happen, but actually (b) is what happens for the majority of initial environment states? I can see that this would be a sensible thing to do if you think of ##\rho^E## as representing an ensemble of initial environments rather than the "true" initial environment.
There is also the separate claim that, if the environment starts in a pure state (a possibility which TI denies but many other interpretations don't) then the linearity alone should leave the whole universe in a superposition of "what would have happened if ##\rho^S## were ##\psi_1##" and the same with ##\psi_2##, which feels to me like it ought to imply an outcome like (a), and it seems like I could then extract (a) for an arbitrary ##\rho^E## by writing it as a convex combination of pure states. I assume this paragraph also contains a mistake, but I would be interested to know where it is.