A Evaluate this paper on the derivation of the Born rule

  • #51
There really is no such thing as collapse. It is as meaningless as saying that the probability P = 1/2 "collapses" once the coin lands heads.

"Probability amplitudes, when squared, give the probability of a complete event. Keeping this principle in mind should help the student avoid being confused by things such as the 'collapse of the wave function' and similar magic" - Richard Feynman
 
Physics news on Phys.org
  • #52
Ddddx said:
There really is no such thing as collapse. It is as meaningless as saying that the probability P = 1/2 "collapses" once the coin lands heads.

The first sentence might be true, but it's not obvious that the second sentence is. That's really the whole point of Bell's theorem. We don't worry about collapse with coin flips because we assume that after a coin settles down, there is a "fact of the matter" about whether it is heads or tails. So even if we flip the coin with the lights out, and don't see the result, we believe that there is a result, we just don't know what it is. The probabilities reflect our ignorance about what that result its.

In the case of quantum mechanics, if you have a particle that is in a superposition of spin-up and spin-down, it's not that it's either spin-up or spin-down, we just don't know which. It's neither spin-up nor spin-down until we measure the spin.
 
  • Like
Likes Zafa Pi
  • #53
Prathyush said:
Fair enough. I use the same formalism.

First a minor point.

I usually don't include the ##\beta## into the definition of the projection operator or born rule, for instance I would write ##\hat{P}_a= |a,\beta \rangle \langle a,\beta| ## and other formulas would change appropriately. Ofcourse it has to be used based on the context.
That's wrong, because then your probabilities won't sum up to 1 (except if the spectrum of ##\hat{A}## is not degenerate, and each eigenspace is exactly 1D).
This is something you are suggesting that is orthogonal to what most textbooks write. I can understand where you are coming from when you say what happens to the system after measurement is irrelevant, I will critically analyze this statement in a later post.

For now however Consider a 2 slit experiment(for instance feynman's description of it), If we measure which slit the particles went through, the for all future measurements we have to use this information. This is the reason why most textbooks include the collapse postulate. When a measurement is performed this information must be reflected in the state of the particle atleast in this particular context. Ofcourse you can say once the measurement is performed we can move both the particles to the same place(or change it however you want), so the collapse posulate is not a general rule. I will analyze this "rule" carefully based on your response.

We can talk about macroscopic and microscopic descriptions once we resolve the collapse stuff.
Again, you have to define the measurement done clearly. Then you won't need a collapse hypothesis. It is not enough to say you gained some which-way information in the double-slit experiment, but you have to say how you definitely measure it to describe the setup of your experiment completely. Then you can say which state is prepared and which probabilities for detecting the particles on the screen you expect from the QT formalism.

One example is to use linearly polarized photons in the double-slit experiment. To gain which-way information you can put quarter-wave plates into each slit, the one in ##+45^{\circ}##, the other ##-45^{\circ}## orientation relative to the polarization direction of the photons. Then a photon running through the first (second) slit will be left-circular the other right-circularly polarized and thus you can exactly distinguish through which way the photons went through the slits. At the same time since the polarization states are exactly perpendicular to each other there is no more interference and thus the interference pattern vanishes. You can also decide to gain incomplete which-way information by distorting the angles of the quarter-wave plates a bit. Then you get partial interference, i.e., an interference pattern with less contrast than without the quarter-wave plates.

The very simple message of this example is that of course the outcome of a measurement depends on the preparation of the measured observable. This is not very profound and is as valid in classical physics as in quantum theory.
 
  • #54
stevendaryl said:
The first sentence might be true, but it's not obvious that the second sentence is. That's really the whole point of Bell's theorem. We don't worry about collapse with coin flips because we assume that after a coin settles down, there is a "fact of the matter" about whether it is heads or tails. So even if we flip the coin with the lights out, and don't see the result, we believe that there is a result, we just don't know what it is. The probabilities reflect our ignorance about what that result its.

In the case of quantum mechanics, if you have a particle that is in a superposition of spin-up and spin-down, it's not that it's either spin-up or spin-down, we just don't know which. It's neither spin-up nor spin-down until we measure the spin.
Indeed, it depends on the preparation whether ##\sigma_z=+1/2## (up) or ##\sigma_z=-1/2## (down) or whether it's indetermined. In the latter case you know, provided you know the quantum state of the measured spin, probabilities for the two possible outcomes. What happens when measuring ##\sigma_z## depends on the used measurement apparatus. There's no general rule like collapse describing what's going on. In my above description of the SG experiment it's clear that you can use the magnetic field to prepare (almost exactly) a pure ##\sigma_z=+1/2## or ##-1/2## state by filtering out the wanted beam since through the magnetic field position and spin get (almost precisely) entangled. But still you don't have a collapse, changing any entity simultaneously everywhere but it's just filtering out one partial beam by blocking the other with some "beam dump". The interaction of these particles with the beam dump is quite local. There's nothing collapsing instaneously in the entire universe as claimed by the collapse proponents.
 
  • #55
stevendaryl said:
In the case of quantum mechanics, if you have a particle that is in a superposition of spin-up and spin-down, it's not that it's either spin-up or spin-down, we just don't know which. It's neither spin-up nor spin-down until we measure the spin.
I don't think your second sentence is quite right. It's true only in the sense that the particle might not even be there at all. Are you saying that Schrodinger's cat is neither alive nor dead until we open the box?

To be pedantic:

An observational apparatus imposes a context -- of a representation (what is to be measured) and an associated frame of reference. We could, in principle, impose those to describe a particle we knew was there without ever observing it. The "probability amplitudes" would have the same meaning in terms of the theoretical asymptotic relative frequencies if a sequence of observations were made. Making one, two or a 100 observations makes no difference to that since we can never make an infinite number of observations. So it is quite possible that the particle was already prepared in a spin eigenstate before we observe it -- as if the notion of "collapse" had already happened when the particle was produced. It's just that we can't start actually measuring those relative frequencies until we start detecting the state and repeating multiple times.

The content of QM then, is not that the particle is not in a spin eigenstate, but that there are multiple representations and frames of reference we could use to describe it before we actually decide what context to use for an actual observation. It is the incompatibility of different contexts that an observer might impose that makes QM differ from classical statistics.
 
  • #56
mikeyork said:
I don't think your second sentence is quite right. It's true only in the sense that the particle might not even be there at all. Are you saying that Schrodinger's cat is neither alive nor dead until we open the box?

I'm saying (actually, I did say) that an electron that is in a superposition of spin-up and spin-down is neither spin-up nor spin-down until we measure it. What this implies about dead cats is complicated.
 
  • Like
Likes vanhees71
  • #57
vanhees71 said:
That's wrong, because then your probabilities won't sum up to 1 (except if the spectrum of ##\hat{A}## is not degenerate, and each eigenspace is exactly 1D).

When I add up probabilities I do a sum over ##\beta##. Its mostly a personal preference, it is equivalent to the formula you wrote. I like my projection operators to have a trace of 1.

Ddddx said:
There really is no such thing as collapse. It is as meaningless as saying that the probability P = 1/2 "collapses" once the coin lands heads.

When measurements are performed new information comes to light and it must be reflected in the new of description of the state. This change in the description as new information becomes available is basically what people call collapse.

vanhees71 said:
There's nothing collapsing instaneously in the entire universe as claimed by the collapse proponents.
This kind of thinking happens because people seem to associate some kind of physical attributes to information, when it does not exist. The wavefunction is basically the same as information available. I want to avoid using the word collapse as it seems to imply things that I don't intend.

vanhees71 said:
Again, you have to define the measurement done clearly. Then you won't need a collapse hypothesis. It is not enough to say you gained some which-way information in the double-slit experiment, but you have to say how you definitely measure it to describe the setup of your experiment completely. Then you can say which state is prepared and which probabilities for detecting the particles on the screen you expect from the QT formalism.

I can construct a detailed experiment, but that would require time. Would you agree with the following statement, when a measurement is performed, the state of the system(meaning information available to us about it) in general changes to reflect the outcome of the measurement.
 
Last edited:
  • #58
stevendaryl said:
I'm saying (actually, I did say) that an electron that is in a superposition of spin-up and spin-down is neither spin-up nor spin-down until we measure it. What this implies about dead cats is complicated.
In the case of the cat, being dead or alive is part of what we mean by it being a cat. So it must be either. In the case of a particle, if we impose a descriptive context that says it must have a definite spin component in a specific direction then that is what it will have (with appropriate probabilities), regardless of whether we measure it.

So I would revise your statement to say "neither spin-up nor spin-down until we choose to describe it as either, whether we measure it or not and if we don't (or can't) choose to describe it as either, then we have nothing to say about it being either". Superpositions tell us only how to switch between incompatible descriptive choices. They don't have any other meaning.

(Edited)
 
Last edited:
  • #59
mikeyork said:
In the case of the cat, being dead or alive is part of what we mean by it being a cat. So it must be either. In the case of a particle, if we impose a descriptive context that says it must have a definite spin component in a specific direction then that is what it will have, regardless of whether we measure it.

So I would revise your statement to say "neither spin-up nor spin-down until we choose to describe it as either, whether we measure it or not and if we don't choose to describe it as either, then we have nothing to say about it being either". Superpositions tell us only how to switch between incompatible descriptive choices. They don't have any other meaning.

I don't think that's true. I should say more definitely: it is not true. Superpositions are not a matter of descriptive choices. To say that an electron is in a superposition \alpha |u\rangle + \beta |d\rangle implies that a measurement of the spin along axis \vec{a} will yield spin-up with a probability given by (mumble..mumble---I could work it out, but I don't feel like it right now). So there is a definite state \alpha |u\rangle + \beta |d\rangle, and it has a definite meaning. It's not just a matter of descriptions.
 
Last edited:
  • Like
Likes edguy99
  • #60
stevendaryl said:
In the case of quantum mechanics, if you have a particle that is in a superposition of spin-up and spin-down, it's not that it's either spin-up or spin-down, we just don't know which. It's neither spin-up nor spin-down until we measure the spin.

Basically when we say we have a particle in a state of superposition, we are saying something about its preparation procedure.

I don't think we should discuss cats here. However for the sake of completeness.

mikeyork said:
I don't think your second sentence is quite right. It's true only in the sense that the particle might not even be there at all. Are you saying that Schrodinger's cat is neither alive nor dead until we open the box?

If you can construct an experiment that can interfere between dead and alive states of a cat you will realize what stevendaryl is saying is correct. However in practice it is impossible to do so.
 
  • #61
Prathyush said:
If you can construct an experiment that can interfere between dead and alive states of a cat you will realize what stevendaryl is saying is correct. However in practice it is impossible to do so.
I disagree. I would claim that the the "probability amplitudes" have exactly the same meaning (an abstract probability that relates to the relative frequencies over an infinite number of identical experiments) whether you make 1, 10, 100 or 0 measurements, with electrons or cats.

In the unfortunate language of "collapse" I am saying that it is just as accurate (or no less inaccurate!) to say it takes place at production as at detection.
 
Last edited:
  • #62
mikeyork said:
I disagree. I would claim that the the "probability amplitudes" have exactly the same meaning (an abstract probability that relates to the relative frequencies over an infinite number of identical experiments) whether you make 1, 10, 100 or 0 measurements, with electrons or cats.

In the unfortunate language of "collapse" I am saying that it is just as accurate (or no less inaccurate!) to say it takes place at production as at detection.
I don't understand at all what you are saying, Collapse is not a physical process. Wavefunction is our description of the system, that is all.
 
  • #63
Prathyush said:
I don't understand at all what you are saying, Collapse is not a physical process. Wavefunction is our description of the system, that is all.
I am saying that probability amplitudes have the same meaning whether any measurements are made or not. To say that spin has certain probabilities of being up or down is not the same as saying it is neither.
 
  • #64
mikeyork said:
I am saying that probability amplitudes have the same meaning whether any measurements are made or not. To say that spin has certain probabilities of being up or down is not the same as saying it is neither.
Probability amplitudes when squared talk about the probabilities of measurements. That is the only way we can use them. You may disagree, but if you want to discuss this point please start a separate thread.
 
  • #65
Prathyush said:
Probability amplitudes when squared talk about the probabilities of measurements. That is the only way we can use them. You may disagree, but if you want to discuss this point please start a separate thread.
No, I don't disagree at all. You just don't have to make a measurement for them to have that meaning.
 
  • #66
The word "collapse" was never used by the founders of quantum theory.

If you look at Feynman's lectures on physics volume 3, you will find exactly zero mentions of that word.

It just isn't proper terminology, and seems to stem from a misunderstanding of what the wave function is.
 
  • #67
Ddddx said:
The idea of "collapse" was never used by the founders of quantum theory.

If you look at Feynman's lectures on physics volume 3, you will find exactly zero mentions of that word.

It just isn't proper terminology, and seems to stem from a misunderstanding of what the wave function is.

The word collapse should not be used. It should simply be called measurement.
 
  • #68
Prathyush said:
I have encountered this paper "Curie Wiess model of the quantum measurement process". https://arxiv.org/abs/cond-mat/0203460

Another work by the same authors is "Understanding quantum measurement from the solution of dynamical models" https://arxiv.org/abs/1107.2138

I am still evaluating the papers. I find the general lessons implied to be interesting and probably compelling.

In the context of the model studied is this paper accurate? What do you think about the overarching viewpoint presented by the authors?
I mentioned this work several time at physicsforums (see, e.g., https://www.physicsforums.com/threa...-local-realism-ruled-out.689717/#post-4372139 )

I believe this is outstanding work, although I cannot check their calculations. I would emphasize the following: 1. They show that the Born rule can be derived from unitary evolution as an approximate, rather than a precise result; 2. The contradiction between unitary evolution and definite outcomes of measurements can be overcome to some extent: the reversal of definite outcomes takes a very large time (Poincare reversal time).
 
  • #69
vanhees71 said:
You can load QT (as any mathematical model of reality) with some philosophical (not to call it esoterical) questions like, why we always measure eigenvalues of self-adjoint operators, but physics doesn't answer why a mathematical model works, it just tries to find through an interplay between measurements and mathematical reasoning such models that describe nature (or even more carefully formulated what we observe/measure in nature) as good as possible.

The purpose of my investigation is to understand the mechanics of measurement, why do measurement apparatus do what they appear to do. Consider a cloud chamber, we understand exactly how it is constructed. Take water molecules do so and so things to it, and we can construct it. We know that upon the interaction with a charged it turns cloudy and in turn we obtain information about its position. Now I want to understand exactly why this happens. Clearly the situation involves the need to describe the cloud chamber using statistical ensembles. The location of the cloud is related to the location of the charged particle. However water molecules are difficult to describe. Can one distil the essence of such a problem into a model. From such a investigation it seems highly compelling to me that Born's rule can be understood from dynamics.
 
Last edited:
  • #70
stevendaryl said:
I don't think that's true. I should say more definitely: it is not true. Superpositions are not a matter of descriptive choices. To say that an electron is in a superposition \alpha |u\rangle + \beta |d\rangle implies that a measurement of the spin along axis \vec{a} will yield spin-up with a probability given by (mumble..mumble---I could work it out, but I don't feel like it right now). So there is a definite state \alpha |u\rangle + \beta |d\rangle, and it has a definite meaning. It's not just a matter of descriptions.
Your axis \vec{a} is a descriptive choice. The probabilities you get are dependent on that choice. Choose a different axis and you'll get different probabilities.
 
  • #71
stevendaryl said:
I'm saying (actually, I did say) that an electron that is in a superposition of spin-up and spin-down is neither spin-up nor spin-down until we measure it. What this implies about dead cats is complicated.
According to standar QT in this case the spin-##z## component is indetermined, and it's not indetermined, because we don't know it but it's really indetermined. All that is known about ##\sigma_z## are the probabilities to find the two possible values when measuring the observable ##\sigma_z##. That's it. There's no more according to QT.
 
  • #72
Prathyush said:
I can construct a detailed experiment, but that would require time. Would you agree with the following statement, when a measurement is performed, the state of the system(meaning information available to us about it) in general changes to reflect the outcome of the measurement.
In the Schrödinger picture the Statistical Operator changes according to the full Hamiltonian of the system. The Hamiltonian must include the interaction with the complete setup. That's all there is, and it's as well valid in classical physics. The only difference is that in classical physics, where you measure macroscopic observables you can make the influence of the measurement apparatus arbitrarily small. This is not the case when you measure microscopic observables. E.g., to measure the electric field of an electron there's no test charge to do so because any charge is at least as big as the electron charge.
 
  • #73
Prathyush said:
I don't understand at all what you are saying, Collapse is not a physical process. Wavefunction is our description of the system, that is all.
Well, in physics you don't want a self-contradictory description. The collapse hypothesis is incompatible with the very foundations of physics, i.e., the causality structure of relativistic spacetime. So why should you assume such a thing? I don't know since I don't know a single example of a real-world experiment, where this assumption is really needed.
 
  • #74
Prathyush said:
This is clearly a question that I haven't thought about in depth, this discussion was extremely fruitful to me because it brought these issues into the forefront.
Chapters 7-10 in my online book derive everything without assuming Born's rule anywhere.
 
  • #75
Well in Chpt. 8 it's just QT in the ##C^*##-algebra formulation. You don't talk about probabilities but about expectation values etc. So just not mentioning the word "probability" doesn't mean that you don't use probability theory.
 
  • #76
vanhees71 said:
define the expectation value of an observable, represented by a self-adjoint operator ##\hat{A}## by
$$\langle A \rangle_{\rho}=\mathrm{Tr}(\hat{\rho} \hat{A}).$$
This is Born's rule.
How can a mathematical definition say anything about the interpretation?

The formula you quote is just part of shut up and calculate. Interpretation enters only if you want to give the expectation value a meaning in terms of measurement.

Standard quantum theory consists of two parts:
  • (S) the shut up and calculate part, which just derives mathematical consequences of definitions, and
  • (I) the interpretation of the formulas from the shut up and calculate part in terms of the real world.
Calling trace##\rho A## the expectation of ##A## and denoting it by ##\langle A\rangle## belong to (S). All rules and results used in statistical mechanics to deduce consequences from it also belong to (S). Only telling what ##\langle A\rangle## should mean belongs to (I). In particular, the shut up and calculate part gets different interpretations depending on how one interprets ##\langle A\rangle##. As equilibrium thermodynamics shows, an interpretation in terms of an average over real measurements is not warranted for macroscopic systems where usually only a single measurement is made and the averaging becomes vacuous. Instead, the standard interpretation of ##\langle A\rangle## in any textbook of statistical thermodynamics (in particular the famous book by Callen) is to equate it with the measured macroscopic value since this identification (and only this) allows one to deduce equilibrium thermodynamics from statistical physics.

vanhees71 said:
what's measured on macroscopic systems usually are indeed very "coarse-grained observables",
In the derivation of equilibrium thermodynamics from statistical physics, coarse graining is never used.

vanhees71 said:
As far as I know, the definition of POVM measurements relies also on standard quantum theory, and thus on Born's rule (I've read about them in A. Peres, Quantum Theory: Concepts and Methods). It just generalizes "complete measurements" by "incomplete ones". It's not outside the standard rules of quantum theory.

1. The POVM formalism belongs to (S); the interpretation in terms of measurement of physical processes in the Lab belongs to (I). Clearly, Born's rule is only an extremal case of the POVM interpretation.

2. POVM's can be analyzed in terms of Born's rule only by going to a fictitious bigger Hilbert space and defining there a new dynamics. This is not the dynamics that one gets naturally from the given system.

3. Even though it can be derived from Born's rule in the above way, a measurement by means of a POVM is not governed itself by Born's rule. You seem to equate everything that can somehow be derived from Born's rule wit Born's rule itself. But this is a severe abuse of language.

For example, homodyne photon measurement measures both the frequency and the phase of a photon, though both are noncommuting variables. This has nothing at all to do with the kind of measurement following the Born rule.
 
Last edited:
  • #77
vanhees71 said:
Well, usual statistics textbook start with equilibrium distributions and define, e.g., the grand canonical operator
$$\hat{\rho}=\frac{1}{Z} \exp[-\beta (\hat{H}-\mu \hat{Q})],$$
to evaluate expectation values using Born's rule, leading to the Fermi- and Bose-distribution functions.

No. these expectation values (of macroscopic quantities) are evaluated with shut up and calculate, not with Born's rule! The results obtained in equilibrium are then identified with the thermodynamic values measured. Nowhere an interpretation in terms of Born's rule enters.

To apply Born's rule to ##\langle H\rangle##, say, one would have to measure astronomically many spectral lines, then do a numerical analysis to extract the energy levels (doing the calculations with a horrendous number of digits to be able to resolve them reliably), and then perform an average over the astronomically large number of energy levels. This is completely ridiculous.

Instead, only as many measurements are performed as there are thermodynamic degrees of freedom, and these are compared with the formulas obtained by shut up and calculate.

vanhees71 said:
Classical statistical mechanics is also based on probability concepts since Boltzmann & Co. I don't know, which textbooks you have in mind!
Boltzmann worked with an ideal gas, where one can apply statistical reasoning (though not Born's rule) by averaging over independent atoms. But it works only there!

In his famous textbook from 1901 where the grand canonical ensemble was introduced, Gibbs never averages over atoms, but over ensembles of macroscopic systems! He was well aware that his ensembles were fictitious ones, made of imagined copies of the macroscopic system at hand, needed to justify the application of statistical concepts to single cases. At his time, mathematics wasn't yet as abstract as today where one can use any mathematical concept as a tool in quite diverse applications where the same mathematical notion has completely different uses and interpretations as long as the axiomatically defined postulates are satisfied. Thus he had to take recourse to a fictitious average where today just a reference to shut up and calculate suffices.

As physics cannot depend on imagined but unperformed experiments, it is clear that his expectations are not averages over many experiments but refer to the single case at hand.
 
  • #78
Prathyush said:
when a measurement is performed, the state of the system(meaning information available to us about it) in general changes to reflect the outcome of the measurement.
When a measurement is performed, the state of the detector changes to a state encoding the measurement result. What happens to the tiny system depends a lot on what it is and how it is measured; for example when measuring a photon it is usually absorbed and no longer exists after the measurement.

The paper in the OP analyzes a very special situation where the measurement is von-Neumann, so that the state after the measurement is an eigenstate corresponding to the measured eigenvalue.
 
  • Like
Likes dextercioby
  • #79
Ddddx said:
The word "collapse" was never used by the founders of quantum theory.
Von Neumann introduced the concept in 1932 under a different name; he called it state vector reduction. The name is not that relevant. What actually happens is.
 
Last edited:
  • #80
vanhees71 said:
Well in Chpt. 8 it's just QT in the ##C^*##-algebra formulation. You don't talk about probabilities but about expectation values etc. So just not mentioning the word "probability" doesn't mean that you don't use probability theory.
By the same reasoning, not mentioning the words coordinates in abstract differential geometry would not mean that you don't use coordinates. The point is that coordinates are not unique, and the meaning of them depends on making choices. Thus not using coordinates is a virtue, and really means that no coordinates are used.

Similarly, there is no probability theory involved in C^*-algebras - nowhere in any definition or result. Probability is not defined until you choose a representation in a separable Hilbert space and an orthonormal basis in it, and it is basis-dependent, whereas the C^*-algebra approach is basis independent. Which choice is the correct one is one of the problems making up the measurement problem that you so despise. But for statistical mechanics one never needs to make a choice of basis as the results are all basis independent. So probability never enters.
 
  • #81
A. Neumaier said:
No. these expectation values (of macroscopic quantities) are evaluated with shut up and calculate, not with Born's rule! The results obtained in equilibrium are then identified with the thermodynamic values measured. Nowhere an interpretation in terms of Born's rule enters.

To apply Born's rule to ##\langle H\rangle##, say, one would have to measure astronomically many spectral lines, then do a numerical analysis to extract the energy levels (doing the calculations with a horrendous number of digits to be able to resolve them reliably), and then perform an average over the astronomically large number of energy levels. This is completely ridiculous.

Instead, only as many measurements are performed as there are thermodynamic degrees of freedom, and these are compared with the formulas obtained by shut up and calculate.Boltzmann worked with an ideal gas, where one can apply statistical reasoning (though not Born's rule) by averaging over independent atoms. But it works only there!

In his famous textbook from 1901 where the grand canonical ensemble was introduced, Gibbs never averages over atoms, but over ensembles of macroscopic systems! He was well aware that his ensembles were fictitious ones, made of imagined copies of the macroscopic system at hand, needed to justify the application of statistical concepts to single cases. At his time, mathematics wasn't yet as abstract as today where one can use any mathematical concept as a tool in quite diverse applications where the same mathematical notion has completely different uses and interpretations as long as the axiomatically defined postulates are satisfied. Thus he had to take recourse to a fictitious average where today just a reference to shut up and calculate suffices.

As physics cannot depend on imagined but unperformed experiments, it is clear that his expectations are not averages over many experiments but refer to the single case at hand.
I think that's just fight about semantics, what you call Born's rule. For me it's the probabilistic interpretation of the state. Usually it's formulated for pure states and then argued for the more general case of mixed states. The upshot is that you can describe the state as a statistical operator ##\hat{\rho}## with the meaning in terms of probabilities given in one of my postings above:
$$P(a)=\sum_{\beta} \langle a,\beta|\hat{\rho}|a,\beta \rangle.$$
That you can identify ##\langle A \rangle## with "the measured value" for macroscopic systems is due to the fact that ##\langle A \rangle## is an observable like, e.g., the center of mass position of some object or a fluid cell and that such observables tend to be sharply peaked around the average value. Of course, a single measurement doesn't tell anything, as everybody learns in the introductory lab in any physics curriculum. What you call "measurement" is indeed not formalized in theory but determined by concrete experimental setups in the lab and real-world measurement devices like detectors.

Of course, you can extend this ideal picture of precisely measured quantities also for microscopic observables with the more general case of incomplete measurements which is formalized as the POVM formalism, but that's finally also based on the fundamental postulates, including Born's rule (at least in the way it's introduced by Peres in his book).
 
  • #82
A. Neumaier said:
By the same reasoning, not mentioning the words coordinates in abstract differential geometry would not mean that you don't use coordinates. The point is that coordinates are not unique, and the meaning of them depends on making choices. Thus not using coordinates is a virtue, and really means that no coordinates are used.

Similarly, there is no probability theory involved in C^*-algebras - nowhere in any definition or result. Probability is not defined until you choose a representation in a separable Hilbert space and an orthonormal basis in it, and it is basis-dependent, whereas the C^*-algebra approach is basis independent. Which choice is the correct one is one of the problems making up the measurement problem that you so despise. But for statistical mechanics one never needs to make a choice of basis as the results are all basis independent. So probability never enters.
Well, we talk about physics not pure mathematics, and you need a rule (called "interpretation") to relate your formalism to what's measured in the real world. This is done by deciding which observable you measure and this determines the basis you have to use to calculate the corresponding probabilities. The ##C^*## formalism is, as far as I can see, equivalent to the standard definition of QT with the advantage to give a more clear mathematical determination of the operator algebra.
 
  • #83
vanhees71 said:
I think that's just fight about semantics, what you call Born's rule. For me it's the probabilistic interpretation of the state.
Then you should say the latter whenever you want to say the former, or you will definitely earn misunderstanding. For the two are far from synonymous. The standard semantics is the one described by wikipedia; nobody apart from you has this far too general usage you just announced. In particular, equating the two is meaningless in the present context - this thread is about deriving Born's rule from statistical mechanics, not about deriving the probabilistic interpretation of quantum mechanics.

vanhees71 said:
Well, we talk about physics not pure mathematics, and you need a rule (called "interpretation") to relate your formalism to what's measured in the real world. This is done by deciding which observable you measure and this determines the basis you have to use to calculate the corresponding probabilities.
Which observable is measured in homodyme photon detection, the example mentioned before?

Moreover, your recipe (calculate the corresponding probability) only works in simple cases where you have an exactly solvable system, hence can evaluate the partition function as a sum over joint eigenstates. But the latter is just one possible way of organizing the computations (shut up and calculate - no interpretation is needed to express the trace as a sum over eigenvalues) and fails in more complex situations.

In equilibrium thermodynamics one wants to measure the total mass of each chemical component (which may be a complex molecule) and the total energy of a macroscopic interacting system. In these cases on never calculates the thermodynamic equation of state in terms of probabilities. Instead one uses mean field approximations and expansions beyond, as you know very well!

In general, a partition sum is just a piece of shut uup and calculate, as it is a mathematically defined expression valid without any interpretation. The interpretation is about relating the final results (the equation of state) to experiments, and this does not involve probabilities at all; it is done simply by equating the expectation of a macroscopic variable with the measured value. Thus this is the true interpretation rule used for macroscopic measurement. Everything else (talk about probabilities, Born's rule, etc.) doesn't enter the game anywhere (unless you want to complicate things unnecessarily, which is against one of the basic scientific principles called Ockham's razor).
 
Last edited:
  • #84
vanhees71 said:
The C^* formalism is, as far as I can see, equivalent to the standard definition of QT with the advantage to give a more clear mathematical determination of the operator algebra.
It is equivalent on the shut up and calculate level, but has a definite advantage of clearness not only on the conceptual but also on the interpretational level. It dispenses with Born's rule, the philosophically problematic concept of probability, and the choice of basis, except when a concrete experiment singles out a concrete basis.

Another advantage is that it directly works with mixed states, which are the by far most common states in Nature, and avoids its decomposition
vanhees71 said:
The upshot is that you can describe the state as a statistical operator ##\hat{\rho}## with the meaning in terms of probabilities given in one of my postings above:
$$P(a)=\sum_{\beta} \langle a,\beta|\hat{\rho}|a,\beta \rangle.$$
which is completely unphysical since the pieces in the sum are far from unique and therefore cannot have a physical interpretation. Different decompositions are physically undistinguishable!
 
Last edited:
  • #85
A. Neumaier said:
I refereed the paper in question here.
I added several paragraphs to the review, summarizing what was actually derived, and pointing out the declared interpretative assumptions of the authors of the paper mentioned in the OP. These assumptions were made explicit in a much later paper, namely:

A.E. Allahverdyan, R. Balian and T.M. Nieuwenhuizen,
A sub-ensemble theory of ideal quantum measurement processes,
Annals of Physics 376 (2017): 324-352.
https://arxiv.org/abs/1303.7257

The traditional motivational introduction to statistical mechanics is based on mixing an ensemble of pure states. However, none of the results of statistical mechanics depends on this motivational prelude. Instead of assuming statistical mechanics without spelling out the precise assumptions made (as in the paper mentioned in the OP and in the big 200 page article with all formal details) - which might suggest that their derivation depends on the traditional foundations -, the authors are here far more explicit about the assumptions necessary to get their results.

They take in this quite recent paper statistical mechanics as a purely formal theory (i.e., in the shut up and calculate mode) and then give new interpretational principles for how this formalism is to be interpreted. In particular, their interpretational principles are independent of Born's rule (as a statement about measurement). As a consequence, the derivation of Born's rule is a result, not a silent assumption. For the present discussion, the most relevant statements from this paper are (emphasis by the authors, but notation for the density operator adapted to the present context):

Allahverdyan Balian and Nieuwenhuizen said:
One should therefore, as done for q-bits, distinguish ##\langle\hat O\rangle=##tr##\rho\hat O## from an ordinary expectation value by denominating it as a
``q-expectation value''. Likewise, a ``q-correlation'', the q-expectation value of a product of two observables, should not be confused with an
ordinary correlation. Also, the q-expectation value ##\langle\hat\pi\rangle## of a projection operator ##\hat\pi## is not an ordinary probability, but a formal object which we will call ``q-probability'' rather than ``probability''. Born's rule is not postulated here, it will come out (Subsec. 6.4) as a property of the apparatus at the issue of an ideal measurement.

Allahverdyan Balian and Nieuwenhuizen said:
Interpretative principle 1. If the q-variance of a macroscopic observable is negligible in relative size its q-expectation value is identified with the value of the corresponding macroscopic physical variable, even for an individual system.

These statements exactly match the assumptions made in my thermal interpretation of quantum mechanics.

By the way, they cite Bell's theorem as a main reason why one cannot simply equate the q-expectations with expectation values in the classical sense since some of the properties of expectations valid in the classical case fail to hold in the quantum case.
 
Last edited:
  • #86
A. Neumaier said:
How can a mathematical definition say anything about the interpretation?
This is an interesting question. Most physicists seem to think it does say everything about the interpretation, in other words the formalism is not neutral as the "shut up and calculate" lemma seems to imply.
By your posts I infer that what you think must inform the interpretation is the macroscopic measurement, but this thread's discussion seems to go in circles, because the formalsm doesn't seem to distinguish clearly between uncertainties derived from classical lack of knowledge and from inherent impossiblity of the theory. So the important distinction between statistics and probabilities that has been made here cannot be resolved by the formalism. But just going to the statistical mechanics interpretation seems to lack a new definition of measurement, or I can't see the improvement over the basis-dependent probabilities, how is the macroscopic measurement of the specific experiment or observation connected to the formalism in a basis-independent way not relyng on the Born rule?.

Also a starting point here seems to be that the Born's rule is just a postulate about probabilities, not acknowledging that the key feature of the rule is that there is a differentiating element with respect to the usual probabilities that is also passed over by the formalism.
 
  • #87
RockyMarciano said:
the formalsm doesn't seem to distinguish clearly between uncertainties derived from classical lack of knowledge and from inherent impossiblity of the theory.
The quantum formalism is independent of knowledge. Subjective issues have no place in physics, except for judging the adequacy of the assumptions and approximations made.
RockyMarciano said:
But just going to the statistical mechanics interpretation seems to lack a new definition of measurement
A measurement of a microscopic system is a reading from a macroscopic device that contains information about the state of the microscopic system. The nature of the coupling and the dynamical analysis must tell which information is encoded in the measurement result, to which accuracy, and with which probabilities.

This definition of a measurement is operationally checkable since one can prepare the states and read the measurement results and can thus compare the theory with the calculations without any ambiguity of concepts.

The only interpretation needed is how the reading from the macroscopic device is related to its macroscopic properties. In the thermal interpretation, this poses no problem at all. The consequences for the microscopic theory are then a matter of deduction, not one of postulation.

Whereas Born's rule is very incomplete in that it doesn't say the slightest thing about what constitutes a measurement, so it is an uncheckable piece of philosophy not of science, unless you know already what measurement means. But this requires knowing a lot of quantum physics that goes into building high quality measurement devices for quantum objects. Thus foundations based on Born's rule are highly circular - unlike foundations based on a properly understood statistical mechanics approach.
 
  • Like
Likes Mentz114
  • #88
A. Neumaier said:
Then you should say the latter whenever you want to say the former, or you will definitely earn misunderstanding. For the two are far from synonymous. The standard semantics is the one described by wikipedia; nobody apart from you has this far too general usage you just announced. In particular, equating the two is meaningless in the present context - this thread is about deriving Born's rule from statistical mechanics, not about deriving the probabilistic interpretation of quantum mechanics.Which observable is measured in homodyme photon detection, the example mentioned before?

Moreover, your recipe (calculate the corresponding probability) only works in simple cases where you have an exactly solvable system, hence can evaluate the partition function as a sum over joint eigenstates. But the latter is just one possible way of organizing the computations (shut up and calculate - no interpretation is needed to express the trace as a sum over eigenvalues) and fails in more complex situations.

In equilibrium thermodynamics one wants to measure the total mass of each chemical component (which may be a complex molecule) and the total energy of a macroscopic interacting system. In these cases on never calculates the thermodynamic equation of state in terms of probabilities. Instead one uses mean field approximations and expansions beyond, as you know very well!

In general, a partition sum is just a piece of shut uup and calculate, as it is a mathematically defined expression valid without any interpretation. The interpretation is about relating the final results (the equation of state) to experiments, and this does not involve probabilities at all; it is done simply by equating the expectation of a macroscopic variable with the measured value. Thus this is the true interpretation rule used for macroscopic measurement. Everything else (talk about probabilities, Born's rule, etc.) doesn't enter the game anywhere (unless you want to complicate things unnecessarily, which is against one of the basic scientific principles called Ockham's razor).
In the Wikipedia article in the first few lines they give precisely the definition, I gave some postings above. I'm using the standard terminology, while you prefer to deviate from it so that we have to clarify semantics instead of discussing physics.

In homodyne detection what's measured are intensities as in any quantum-optical measurement. I refer to Scully&Zubarai, Quantum Optics. One application is to characterize an input signal (em. radiation) (annihilation operator ##\hat{a}##) using a reference signal ("local oscillator", annihilation operator ##\hat{b}##). They are sent through a beam splitter with transmittivity ##T## and reflectivity ##R##, ##T+R=1##. The states at the two output channels are then defined by (I don't put hats on top of the operators from now on):
$$c=\sqrt{T} a + \mathrm{i} \sqrt{1-T} b, \quad d=\sqrt{1-T}a + \sqrt{T} b.$$
What's measured is the intensity at channel ##c##, i.e., ##c c^{\dagger}##.

If the local oscillator is in a coherent state ##|\beta_l \rangle## you get for the expectation value
$$\langle c^{\dagger} c \rangle=T \langle a^{\dagger} a \rangle + (1-T)|\beta_l|^2 - 2 \sqrt{T(1-T)} |\beta_l| \langle X(\phi_l+\pi/2)$$
with
$$X(\phi)=\frac{1}{2} (a \exp(-\mathrm{i} \phi)+a^{\dagger} \exp(\mathrm{i} \phi).$$
All this is done within standard QT using Born's rule in the above given sense. I don't see, which point you want to make with this example. It's all standard Q(F)T.

Now you switch to partition sums, i.e., thermodynamical systems. Take as an example black-body radiation (or any other ideal gas of quanta), i.e., a radiation field in thermal equilibrium with the walls of a cavity at temperature ##T=1/\beta##.
The statistical operator is
$$\hat{\rho}=\frac{1}{Z} \exp(-\beta \hat{H}).$$
The partition sum here is
$$Z=\mathrm{Tr} \exp(-\beta \hat{H}).$$
The Hamiltonian is given by (I use a (large) quantization volume ##V## with periodic boundary conditions for simplicity)
$$\hat{H}=\sum_{\vec{n} \in \mathbb{Z}^3, h \in \{\pm 1\}} \omega(\vec{p}) \hat{a}^{\dagger}(\vec{p},h) \hat{a}(\vec{p},h), \quad \vec{p} = \frac{2 \pi}{L} \vec{n}.$$
For the following it's convenient to evaluate the somewhat generalized partition function
$$Z=\mathrm{Tr} \exp(-\sum_{\vec{n},h} \beta(\vec{n},\lambda) \omega_{\vec{n}} \hat{N}(\vec{n},h).$$
Using the Fock states leads to
$$Z=\prod_{\vec{n},h} \frac{1}{1-\exp(-\omega_{\vec{n}} \beta(\vec{n},\lambda))}.$$
The thermodynamic limit is given by making the volume ##V=L^3## large:
$$\ln Z=-\sum_{\vec{n},h} \ln [1-\exp(-\omega_{\vec{n}} \beta(\vec{n},\lambda)]=-V \sum_{h} \int_{\mathbb{R}^3} \frac{\mathrm{d}^3 \vec{p}}{(2\pi)^3} \ln [1-\exp(-\beta(\vec{p},h) |\vec{p}|].$$
The spectrum (i.e., the mean number of photons per three-momentum) is calculated by
$$\langle N(\vec{p},h) \rangle=-\frac{1}{\omega_{\vec{p}}} \frac{\delta}{\delta \beta(\vec{p},h)} \ln Z=\frac{V}{\exp(\beta |\vec{p}|)-1}.$$
It's measured with help of a spectrometer (or with the Planck satellite for the cosmic microwave background).

It's all standard QT and uses, of course, Born's rule.
 
  • #89
A. Neumaier said:
Whereas Born's rule is very incomplete in that it doesn't say the slightest thing about what constitutes a measurement, so it is an uncheckable piece of philosophy not of science, unless you know already what measurement means. But this requires knowing a lot of quantum physics that goes into building high quality measurement devices for quantum objects. Thus foundations based on Born's rule are highly circular - unlike foundations based on a properly understood statistical mechanics approach.
No theory (also not classical mechanics or field theory) say "the slightest thing about what constitutes a measurement". Physical observables are defined by concrete measurement devices in the lab, not by a theoretical formalism. The theoretical formalism rather gives a mathematical description of such observations. As the name already tells, a statistical mechanics (or rather physics) approach, also uses probabilities in its foundations, or what else is statistics than applied probability theory?

Only with Born's rule the quantum formalism gets interpretible without contradictions with experience. It's not enough to give the other postulates (concerning the formal math describing quantum kinematics and dynamics).
 
  • #90
vanhees71 said:
In the Wikipedia article in the first few lines they give precisely the definition, I gave some postings above. I'm using the standard terminology, while you prefer to deviate from it so that we have to clarify semantics instead of discussing physics.
Well, when the concepts are not clear one must first clarify the semantics before one can communicate physics.

You give conflicting definitions of what you mean by the Born rule, but not all can be true. For example you said in post #27 that the definition ##\langle A\rangle =##tr ##\rho A## is Born's rule. Where in the first few lines of the Wikipedia article is this stated?
 
  • #91
vanhees71 said:
Physical observables are defined by concrete measurement devices in the lab, not by a theoretical formalism.
These concrete devices are calibrated by using quantum mechanical theory for checking that they actually do what they do. Without having already quantum mechanics working one cannot validate any of these checks. One doesn't know the state a laser produces without knowing the theory of the laser, etc. Thus one cannot check the definition of a physical observable (such as spin up) that goes into the theory with which something is computed without having already the theory. This is standard circularity.
 
  • #92
vanhees71 said:
Now you switch to partition sums, i.e., thermodynamical systems.
and you choose an exactly solvable system, which I said are very special cases, the only cases where one can use the sum over probabilities to calculate the partition function. Yes, in particular cases, Born's rule applies and probabilities are used to do the calculations. But these are very special cases.

And even in your partition sum there is not a single measurement but only computations, hence Born's rule (which, according to wikipedia, is ''a law of quantum mechanics giving the probability that a measurement on a quantum system will yield a given result'') doesn't apply. You pay lip service to Born's rule but you don't use it in your computations.
 
  • #93
vanhees71 said:
In homodyne detection what's measured are intensities
What I meant was using homodyne detection to measure simultaneously the quadratures (which are noncommuting optical analogues of position and momentum) by splitting the photon beam 50:50 and then making homodyne measurements on each beam. Of course the raw measurements are measurements of intensities, but in terms of the input, what is measured (inaccurately, within the validity of the uncertainty relation) are noncommuting quadratures.
 
Last edited:
  • #94
vanhees71 said:
Physical observables are defined by concrete measurement devices in the lab, not by a theoretical formalism.
Take as very concrete example the Hamiltonian, which is the observable that goes into the computation of the partition function of a canonical ensemble. How do you define this observable by a concrete measurement device in the lab, that would give according to Born's rule as measurement result the ##k##th energy level ##E_k## with probability ##e^{-\beta E_k}##?

The impossibility of giving such a device proves that defining the meaning of observables and of (accurate) measurements is a thoroughly theoretical process, not just one of experimentation!
 
  • #95
A. Neumaier said:
These concrete devices are calibrated by using quantum mechanical theory for checking that they actually do what they do. Without having already quantum mechanics working one cannot validate any of these checks. One doesn't know the state a laser produces without knowing the theory of the laser, etc. Thus one cannot check the definition of a physical observable (such as spin up) that goes into the theory with which something is computed without having already the theory. This is standard circularity.
Sure, it's well known that physics is "circular" in this way. You need theory to construct measurement devices. At the same time these devices are used to check the very theory on which there construction is based. In a sense, testing the theories is just a test of the consistency of the theory with the observations made.

Spin is a good example. The Stern Gerlach experiment was undertaken before quantum theory in its modern form and before also the modern notion of "spin" has been discovered. The theory used was classical mechanics with some ideas from the Bohr-Sommerfeld model and what was tested were hypotheses based on it. The main trouble in this context was the "anomalous Zeeman effect" which could not be well explained by the Bohr-Sommerfeld model. For a very amusing account of the history (including the fact that without bad cigars the SG experiment most likely would have failed ;-)), see

https://faculty.chemistry.harvard.edu/files/dudley-herschbach/files/how_a_bad_cigar_0.pdf
 
  • Like
Likes Leo1233783
  • #96
A. Neumaier said:
What I meant was using homodyne detection to measure simultaneously the quadratures (which are noncommuting optical analogues of position and momentum) by splitting the photon beam 50:50 and then making homodyne measurements on each beam. Of course the raw measurements are measurements of intensities, but in terms of the input, what is measured (inaccurately, within the validity of the uncertainty relation) are noncommuting quadratures.
I'm not familiar with all applications of homodyne measurements. Before I can comment on this, please give a definition of what experiment you precisely have in mind. What's measured in Quantum Optics are, technically speaking, usually correlation functions of field operators. Such correlation functions sometimes refer to "incomplete measurements" of incompatible observables. How does this, in your opinion, invalidate the standard postulates of Q(F)T? I'm not aware that quantum optics is based on another theory than standard QT.
 
  • #97
A. Neumaier said:
Take as very concrete example the Hamiltonian, which is the observable that goes into the computation of the partition function of a canonical ensemble. How do you define this observable by a concrete measurement device in the lab, that would give according to Born's rule as measurement result the ##k##th energy level ##E_k## with probability ##e^{-\beta E_k}##?

The impossibility of giving such a device proves that defining the meaning of observables and of (accurate) measurements is a thoroughly theoretical process, not just one of experimentation!
Hm, that's not so easy. In principle you can measure it by looking at the emission spectrum of the gas (of course the temperature should be large enough so that the higher than ground states are populated). The relative strengths of different lines is governed by the Boltzmann distribution.
 
  • #98
vanhees71 said:
it's well known that physics is "circular" in this way.
But theoretical physics does not need to be circular; one can have a good theory with a noncircular interpretation in terms of experiments.

While one is still learning about the phenomena in a new theory, circularity is unavoidable. But once things are known for some time (and quantum physics is known in this sense for a very long time) the theory becomes the foundation and physical equipment and experiments are tested for quality by how well they match the theory. Even the definitions of units have been adapted repeatedly to better match theory!

vanhees71 said:
Hm, that's not so easy. In principle you can measure it by looking at the emission spectrum of the gas
But this gives you energy differences, not energy levels. This does not even closely resemble Born's rule!
Moreover, it is a highly nontrivial problem in spectroscopy to deduce from a collection of measured spectral lines the energy levels! And it cannot be done for large molecules over an extended energy range, let alone for a brick of iron.

vanhees71 said:
The relative strengths of different lines is governed by the Boltzmann distribution.
No. It depends also on selection rules and how much they are violated in a particular case. It is quite complicated.

vanhees71 said:
I'm not familiar with all applications of homodyne measurements. Before I can comment on this, please give a definition of what experiment you precisely have in mind.
I mentioned everything necessary. To approximately measure the two quadratures of photons in a beam one passes them through a symmetric beam splitter and then measured the resulting superposition of photons in the two beams by a homodyne detection on each beam. Details are for example in a nice little book by Ulf Leonhardt, Measuring the quantum state of light. This is used in quantum tomography; the link contains context and how the homodyne detection enters.
 
Last edited:
  • #99
vanhees71 said:
Well, in physics you don't want a self-contradictory description. The collapse hypothesis is incompatible with the very foundations of physics, i.e., the causality structure of relativistic spacetime. So why should you assume such a thing? I don't know since I don't know a single example of a real-world experiment, where this assumption is really needed.

I won't use the world collapse form now on, It has meanings that I don't intend. It is also very bad terminology. Let's use the following language from now on, We prepare a system in a state, described as ##|\psi_{in}>##. The system was measured to be in a state described as ##|\psi_{out}>## with a probability ##<\psi_{in}|\psi_{out}>^2##, When we use apparatus where we destroy the particle the appropriate clarification must be made. The wave function is our description of the system. What ##|\psi_{in}>## and ##|\psi_{out}>## are depend on the details of the experimental apparatus.

This must be non controversial to both of us.(Right?)
 
Last edited:
  • #100
Prathyush said:
I won't use the world collapse form now on, It has meanings that I don't intend. It is also very bad terminology. Let's use the following language from now on, We prepare a system in a state, described as ##|\psi_{in}>##. The system was measured to be in a state described as ##|\psi_{out}>## with a probability ##<\psi_{in}|\psi_{out}>^2##, When we use apparatus where we destroy the particle the appropriate clarification must be made. The wave function is our description of the system. What ##|\psi_{in}>## and ##|\psi_{out}>## are depend on the details of the experimental apparatus.
That's not how I would describe things. First off, I would not use the term "measured"; I would rather refer to "state preparation" and "state detection". In the case of detection, it is an eigenvalue of ##a## in representation ##A## chosen by the apparatus that is detected. But we must also take into account the role of the detection apparatus, since the detection process is one of interaction.

The "scattering amplitude" for the interaction is then ##<\psi_i,\phi_i|\psi_f,\phi_f>## where ##\psi_i, \psi_f## are the initial (prepared) and final states of the system that is detected and ##\phi_i, \phi_f## are the initial and final states of the detection apparatus. The detected value ##a## is then interpreted from the change to the apparatus as a function of ##\phi_i## and ##\phi_f## with probability given by the square modulus of the scattering amplitude. In the case that the change in the apparatus is sufficiently small (##\phi_f\approx \phi_i##) and ##\psi_f = \psi_a## is the eigenstate of ##A## with eigenvalue ##a## and then we would have that ##|<\psi_f |\psi_a>|^2## is an approximation to the probability of finding the state ##a##.
 
Last edited:
Back
Top