I The thermal interpretation of quantum physics

  • #251
A. Neumaier said:
But maybe consciousness is not preprogrammed and allows for some user decisions...

Decisions with respect to what in a pre-programmed world?
 
  • Like
Likes AlexCaledin
Physics news on Phys.org
  • #252
akhmeteli said:
Let me remind you the context of the discussion. Previously I wrote:

A. Neumaier replied:

At this point I asked for a reference.

So Peter Morgan explains to me that if we use the anticommutative structure we get a problem with the Dirac equation, but I spoke about "relativistic quantum mechanics, without using quantum field theory." If you use the anticommutative structure and the spectrum is bounded below by the vacuum energy, you actually use quantum field theory.

A. Neumaier explains to me that the Dirac equation is not good enough. But that was exactly my point: if the original nonrelativistic Schrödinger equation and the Dirac equation have problems, how can one fault the Born's rule for that? And again, I have no investment in the Born's rule, I am just trying to say that the specific critique of the Born's rule does not look well-founded.​
@akhmeteli, I read through the whole thread in one sitting after a few weeks of all of you discussing away, so I've no doubt I missed some of the nuances of what each of you hope to get from it. The lack of threading on PF makes it difficult enough to get into a long conversation that I don't usually bother once there are many posts, but I thought that you might find the Hegerfeldt reference of some interest, given that no-one else responded to your request. The connection of positive frequency with analyticity and locality/nonlocality has always seemed to me less emphasized than it might be, particularly given the analogy it suggests with the perhaps simpler connection of the Hilbert transform with analyticity in signal analysis.
In any case, I didn't notice that you wanted to discuss only "relativistic quantum mechanics, without using quantum field theory", about which I know too little to contribute much, but I have always taken it that problems such as you mention with the Dirac equation as relativistic QM are what pushed physicists towards the Dirac equation as the basis of a QFT.
 
  • #253
A. Neumaier said:
No. Like you did in the classical case, I nowhere assume anything probabilistic in the quantum case. Randomness appears as in classical physics by breaking metastability through effects of the deterministic noise neglected in coarse-graining. Of course it will be in agreement with the standard interpretation where the latter is based on actual measurement.
That's what you claim! Your entire formalism is just equivalent to the standard formalism, you just rename "probability" as "q-expectation" and forbid people to translate it back to the standard language which is standard language, because it makes the physical meaning (i.e., the interpretation!) of the formalism clear instead of hiding it for some ununderstandable philosophical quibbles unrelated with physics.
 
  • #254
A. Neumaier said:
It doesn't matter how one calls it. Tradition calls it expectation value and denotes it by pointed brackets, even in situations like:

where it is very clear (and where you agree) that it cannot have this meaning:

Therefore the right way is to regard all these as calculational tools, as you emphasized in this particular case. I call them q-expectations to emphasize that they are always calculational tools and not expectations values in the statistical sense, but using any other name (e.g., ''reference values'', as I did very early in the development of the thermal interpretation) would not change anything.

In some instances (often at the end of long calculations) q-expectations may refer to sample means of actual measurements, but in as many other instances, they refer to single actual measurement only. No matter which interpretation of quantum mechanics is used, the interpretation is done only at this very last stage, namely where actual measurement is involved. There, depending on the situation, one must give them a deterministic (currents, densities) or a statistical (event counts) interpretation.

Thus to account for the actual practice of quantum mechanics in all applications, one needs both deterministic measurements and statistical measurements. The thermal interpretation accounts for both, without having to start with fundamental probability.
As I stressed many times before, the problem is not the formalism, which is anyway leading to the correct results, i.e., the QT results. The problem is that you just rename everything and claim, there's no probability involved just to solve some phyilosophical quibble which are no problems for physics anyway. Then the meaning of the formalism as a physical theory (not an empty mathematical game!) gets lost although that's the very purpose of any interpretation.
 
  • #255
A. Neumaier said:
The average taken is not over microscopic degrees of freedom but over a small spacetime region where the measurement is performed

A. Neumaier said:
sample means of actual measurements, but in as many other instances, they refer to single actual measurement only.

vanhees71 said:
The problem is that you just rename everything and claim, there's no probability involved just to solve some phyilosophical quibble

I am wondering that the TI sort of takes this view based on some idea that I have thought about before. I will explain but Arnold may correct me of course. There is a class of math branch which is called geometric probability( eventually linked to many transforms), while not exactly Schrodinger EQ. type solution (including the density matrix) but a very close cousin that I think illuminates what TI intends . The the class of problems like "line circle picking" or ball triangle picking, and many others that you can check in the links give the probability density function for these picking processes.

Now the density of the line lengths or triangle areas at different values are similar to QM probability density. So what I think TI is trying to say is That these underlying microstates are of no importance in themselves but the mean average is what is important which we actually measure. In this sense Ti denies the physicality of those "random" events and just attributes them to a field like values, and since fields are controversial physical thing, hence the confusion.

My hopeful solution is that these are indeed have physical interpretation(maybe energy since they represent lengths, areas..I don't know) that keeps the interpretations "probabilistic" nature only in the sense that how many line lengths s1 as compared to s2 and so on can fit the area. AND have the TI interpretation as the expectation(average value) translates to the particle "measured" property. I am summering but I hope you get the drift. Of course, I am far from being any certain about these ideas, but at least I hope I am right about my understanding of TI, however general.

P.S. curiously enough many of the pick problems linked to have density functions which look surprisingly similar to what we see in QM :smile:
 
Last edited:
  • #256
Peter Morgan said:
@akhmeteli, I read through the whole thread in one sitting after a few weeks of all of you discussing away, so I've no doubt I missed some of the nuances of what each of you hope to get from it. The lack of threading on PF makes it difficult enough to get into a long conversation that I don't usually bother once there are many posts, but I thought that you might find the Hegerfeldt reference of some interest, given that no-one else responded to your request. The connection of positive frequency with analyticity and locality/nonlocality has always seemed to me less emphasized than it might be, particularly given the analogy it suggests with the perhaps simpler connection of the Hilbert transform with analyticity in signal analysis.
In any case, I didn't notice that you wanted to discuss only "relativistic quantum mechanics, without using quantum field theory", about which I know too little to contribute much, but I have always taken it that problems such as you mention with the Dirac equation as relativistic QM are what pushed physicists towards the Dirac equation as the basis of a QFT.
Of course, nobody can demand that you thoroughly read an entire long thread before you post a reply. And I agree that the Hegerfeldt reference is interesting, but I don't have much to say about it. Hegerfeldt himself offers some plausible interpretations of his results.
 
  • #257
A. Neumaier said:
It is the trace of the product of the q-observable ##j_0(h):=\int_\Omega h(x)j_0(x)dx## with the density operator, ##\langle j(h)\rangle:=Tr~\rho j(h)##, where ##\Omega## is the region in which the coarse-grained current is observed and ##h(x)## is an appropriate smearing function determined by the sensitivity of the measuring instrument or coarse-graining. This is mathematically well-defined; no mystery and no circularity is present (unless you impose your interpretation in addition to mine). The result can be compared with experiment in a single reading, without any statistics involved, giving an operational definition of the meaning of this q-expectation. The average taken is not over microscopic degrees of freedom but over a small spacetime region where the measurement is performed (needed to turn the distribution-valued operator current into a well-defined q-observable,
To the contrary! The average you describe is precisely over "many microscopic degrees of freedom".

Let's first get the notation straight. If I understand it right, your integral is operator valued, i.e., you define a "smeared charge-density operator":
$$\hat{j}_{\Omega}^0(t,\vec{x}_0)=\int_{\Omega} \mathrm{d}^3 x \hat{j}^0(t,\vec{x}).$$
If you want to use this to evaluate a macroscopic charge density, the ##\Omega## must be a "macroscopically small but microscopically large" finite volume around ##\vec{x}_0##.

As important is that the state, ##\hat{\rho}##, allows for such a "separation of scales" (usually a save assumption is that ##\hat{\rho}## is close to thermal equilibrium). Then
$$\rho(t,\vec{x}_0)=\mathrm{Tr} [\hat{\rho} \hat{j}_{\Omega}^0(t,\vec{x}_0)]$$
leads to a physically (!) interpretible "macroscopic, (semi-)classical" charge-distribution observable.

It's important for the physicist to have such an intuitive/heuristic picture to derive usefull quantities. It's not enough to relabel the standard notation with highly abstract formalized language, loosing the contact with the experimental physicist in the lab. Interpretation is not about formal reformulations but precisely about this connection between the formalism and real-world experiments.

Take the other example about the redefinition of several base units of the SI. It may look trivial: You just take a bunch of fundamental natural constants like ##f_{\text{Cs}}##, ##c##, ##h##, ##N_A##, ##k_{\text{B}}##, and define values for them. Formally that's it. With choosing these values and naming the units everything is formally well defined, but that's useless for the physicist, and you have to "realize" the units, based on these definitions in terms of real-world experimental setups. That's what took decades of hard work by the metrologists around the world, not the definition, which is nearly what Planck already has worked out (for technical reasons it's today impossible to do the final step and base everything on Planck units because of the trouble with the determination of Newtons gravitational constant ##\gamma## which could be used to eliminate ##t_{\text{Cs}}## an base everything on fixing fundamental constants rather than to a property of a a certain substance, in this case the atomar finestructure of Cs). It is the realization of the units with the best available technology today to reach the precision necessary to keep the changes to the actual meaning of the units minimal. FAPP there shouldn't be a big change in what 1 kg of mass means. That's why for decades the two approaches to realizing the kg, based on fixing the value of ##h## as in the formal definition, have been pursued by various national metrology institutes (e.g, the PTB with the Avogadro project and NIST with the Kibble balance). The advantage of the redefinition simply is that the fundamental definitions do not change while the ever more precise realization of the units may change with the advance of technology.
 
  • #258
vanhees71 said:
To the contrary! The average you describe is precisely over "many microscopic degrees of freedom".

Let's first get the notation straight. If I understand it right, your integral is operator valued, i.e., you define a "smeared charge-density operator":
$$\hat{j}_{\Omega}^0(t_0,\vec{x}_0)=\frac{1}{|\Omega|}\int_{\Omega} dtd^3\vec{x}\,\hat{j}^0(t,\vec{x}).$$
If you want to use this to evaluate a macroscopic charge density, the ##\Omega## must be a "macroscopically small but microscopically large" finite space-time volume around ##(t_0,\vec{x}_0)##.

As important is that the state, ##\hat{\rho}##, allows for such a "separation of scales" (usually a safe assumption is that ##\hat{\rho}## is close to thermal equilibrium). Then
$$\rho(t,\vec{x}_0)=\mathrm{Tr} [\hat{\rho} \hat{j}_{\Omega}^0(t,\vec{x}_0)]$$
leads to a physically (!) interpretible "macroscopic, (semi-)classical" charge-distribution observable.
I added a time integration to account for the smearing in time necessary for a realistic measurement and the factor ##1/|\Omega|## to make the density correctly normalized. Then this is exactly what I wrote, for the special case of a constant weight function ##h(x)## centered at ##(t_0,\vec{x}_0).##

It involves nothing but an average over a macroscopically small but microscopically large region ##\Omega## of space-time, because that's due to the macroscopic nature of any charge measurement device.

There is no average over microscopic degrees of freedom - these microscopic degrees of freedom appear nowhere in the formula. Nevertheless, one has the correct intuitive/heuristic picture!
 
Last edited:
  • #259
Obviously we have a completely different understanding about the meaning of the symbols. Of course, your formal additions, i.e., using averages over a space-time region rather than only space instead of the sum, are not the point. The point is that the said integral in fact averages and the following tracing with ##\hat{\rho}## averages indeed over many microscopic degrees of freedom.

The classical analogue a la Lorentz is to start with a "microscopic" current
$$j_{\text{mic}}^{\mu}(t,\vec{x})= \sum_N \int_{\mathbb{R}} \mathrm{d} \tau_N q_N \frac{\mathrm{d} y_N^{\mu}}{\mathrm{d} \tau_N} \delta^{(4)}(x-y_N(\tau_N))$$
and then average in the very same sense as above (I'm not sure whether Lorentz averaged over small time intervals too, but that doesn't change the principle of the argument. It's "coarse-graining" in the sense of "blurring out" many microscopic details and fluctuations, introducing a scale of resolution via the extent of the spatial (and temporal) volume (four-volume) elements, ##\Omega##.
 
  • #260
vanhees71 said:
tracing with ##\hat{\rho}## averages indeed over many microscopic degrees of freedom.
No. The trace is a calculational tool only, a purely mathematical operation, namely the sum of the diagonal elements in an arbitrary basis.

Instead of starting with the density operator and the trace, I could as well have defined (as standard in algebraic quantum field theory) ##\langle\cdot\rangle## to be a normalized monotone linear functional - where no trace of a trace exists and the connotation of an average is completely absent. Then one can reconstruct (if desired) the representation in terms of a density operator and the trace by the general machinery of Hilbert space operator theory. But this shows that there is no fundamental averaging going on in the foundations!

You defined Green's functions by ##W(x,y):=Tr~ \rho \phi(x)^*\phi(y)## and though calling them expectation values (which they are not, at least not in the sense of Born's rule), you insisted on that they are only calculational tools. In the same spirit you should allow me to use the trace everywhere in this purely formal meaning as a tool. This does not affect at all the relation to experiment. It just provides everything with a different intuition - a different interpretation of quantum mechanics! In general, to establish a valid relation between the formal tools and reality, only the interpretation of the final predictions in terms of actual measurements counts, not how one calls or thinks of the calculational tools one uses.

The trace is therefore not an average over anything - it is the latter only in your interpretation, and only when the operator is self-adjoint! Your arguments are just a rant against any other interpretation than what you are used to.
vanhees71 said:
The classical analogue a la Lorentz is
... irrelevant for a quantum foundation, since classical physics should be derived from quantum physics and not the other way round.
 
  • Like
Likes Mentz114
  • #261
A. Neumaier said:
No. The trace is a calculational tool only, a purely mathematical operation, namely the sum of the diagonal elements in an arbitrary basis.

Instead of starting with the density operator and the trace, I could as well have defined (as standard in algebraic quantum field theory) ##\langle\cdot\rangle## to be a normalized monotone linear functional - where no trace of a trace exists and the connotation of an average is completely absent. Then one can reconstruct (if desired) the representation in terms of a density operator and the trace by the general machinery of Hilbert space operator theory. But this shows that there is no fundamental averaging going on in the foundations!

You defined Green's functions by ##W(x,y):=Tr~ \rho \phi(x)^*\phi(y)## and though calling them expectation values (which they are not, at least not in the sense of Born's rule), you insisted on that they are only calculational tools. In the same spirit you should allow me to use the trace everywhere in this purely formal meaning as a tool. This does not affect at all the relation to experiment. It just provides everything with a different intuition - a different interpretation of quantum mechanics! In general, to establish a valid relation between the formal tools and reality, only the interpretation of the final predictions in terms of actual measurements counts, not how one calls or thinks of the calculational tools one uses.

The trace is therefore not an average over anything - it is the latter only in your interpretation, and only when the operator is self-adjoint! Your arguments are just a rant against any other interpretation than what you are used to.

... irrelevant for a quantum foundation, since classical physics should be derived from quantum physics and not the other way round.
Averaging is difficult to grasp at the foundational level because it seems that if occupation probabilities are averaged this increases entropy (like replacing numbers ##n_i## with ##\sum n_i/N##) but averaging over dynamical variables decreases entropy ( disorder, uncertainty) which throws a spanner in the search for homologies between probability evolution and thermodynamic evolution.
 
Last edited:
  • #262
I think we discuss in circles, because you don't want to understand my point.

On the one level you have a formalism. There you can use any calculational tool you like to get the physical numbers out. That's a technical point.

However, from a paper claiming to have a new interpretation I expect an interpretation not just the mathematical formalism. The interpretation is precisely what's beyond the mathematical formalism making it from a pure game of human thought as, say, "group theory", to a physical theory. The Green's function is defined as a trace, and that's why I write it in terms of the expecation value ##\langle \hat{\phi}(x) \hat{\phi}^{\dagger}(y) \rangle##. This doesn't imply at all that it is an observable quantity. It's just a useful tool to calculate an observable quantity, e.g., the single-particle phase-space distribution function after precisely using the coarse-graining formalism we discuss right now to its Wigner transform, and it is very clear that the coarse graining is over "many microscopic degrees of freedom", which motivates the entire formalism to begin with. There are different aims for mathematicians and physicists. The mathematician will be satisfied with clarifying the formal sense of the calculational operations physicists use to get their numbers out, but that's not what you expect if you read a paper about interpretation.

I brought the classical example since usually it's easier to understand, in which sense a formal averaging/coarse-graining procedure is about summing/averaging over a huge number of microscopic degrees of freedom. Obviously this was another failed attempt to bring the argument across :-((.
 
  • #263
vanhees71 said:
I think we discuss in circles, because you don't want to understand my point.
No; we discuss in nonintersecting lines, because neither gets his point across. This is a thread about the thermal interpetation, so the rules of the thermal interpretation should be applied, and you should want to understand my point rather than try to make a point based on assuming the statistical interpretation.

vanhees71 said:
from a paper claiming to have a new interpretation I expect an interpretation not just the mathematical formalism. The interpretation is precisely what's beyond the mathematical formalism making it from a pure game of human thought as, say, "group theory", to a physical theory.
I fully agree. But my interpretation is opposite from yours, except at those places where theoretical predictions are compared with event statistics, where they agree.

To interpret a spacetime integral of the theoretical q-expectation ##\langle j_0(x)\rangle## of a density as a single measured value of an instrument with a limited spacetime resolution tells directly and in a very intuitive manner what the theoretical q-expectation means in terms of observability. There is no averaging of actual measurements. And there is no need to talk about any postulated unobservable microscopic averaging, which is nowhere in the mathematical formalism, and nowhere in the calculations.

Similarly, to interpret a field correlation, i.e., a theoretical q-expectation ##\langle \phi(x)^*\phi(y)\rangle## in terms of the integral kernel of an instrument measuring the linear response to an external field perturbation requires a fluctuation-dissipation theorem (a theoretical result on the level of the mathematical formalism) and again tells directly and in a very intuitive manner what the theoretical correlation function means in terms of observability. There is no averaging of actual measurements. And there is no need to talk about any postulated unobservable microscopic averaging, which is nowhere in the mathematical formalism, and nowhere in the calculations.

Similarly, to interpret a spectral density function, i.e., the Fourier transform of another theoretical q-expectation in terms of an instrument measuring the positions and width of spectral lines again tells directly and in a very intuitive manner what the theoretical spectral density function means in terms of observability. There is no averaging of actual measurements. And there is no need to talk about any postulated unobservable microscopic averaging, which is nowhere in the mathematical formalism, and nowhere in the calculations.

The same holds for all other connections to experiment, including those where some statistics is done to get the observations! In the latter case, only the same kind of statistics done in classical measurement theory to go from many highly uncertain actual observations to a much more predictable statistical average. But there is never a need to talk about any postulated unobservable microscopic averaging, which is nowhere in the mathematical formalism, and nowhere in the calculations.

The talk about averaging unobservable stuff has in the thermal interpretation the same dispensable superficial (and often misleading) meaning as the talk about virtual particles explaining how particles mediate interactions between fields. They are an artifact of simplified imagery applied to formal developments in a dubious attempt to make them less abstract and more intuitive. The expert knows in which sense and with how many grains of salt it has to be taken to guide calculations without being led astray, but whoever tries to take the talk at face value is lead into the morass that 93 years of quantum interpretations had not been able to clear up.
 
  • Like
Likes dextercioby
  • #264
vanhees71 said:
it is very clear that the coarse graining is over "many microscopic degrees of freedom"
In the thermal interpretation, the coarse-graining is not done by averaging over many microscopic degrees of freedom (which nowhere figure in my description of coarse-graining in Subsection 4.2 of Part III).

Coarse-graining is done by picking a vector space of relevant q-expectations and restricting the exact dynamics of all q-expectations to an exact or approximate dynamics of the relevant q-expectations, using one of a number of theoretical techniques described purely on the mathematical level. Examples are the Dirac-Frenkel variational principle, the projection operator formalism, or the closed-time-path (CTP) formalism. Where is the averaging over many microscopic degrees of freedom in the CTP formalism as described in the book by Calzetta and Hu? They talk informally about looking at slow variables and ignoring obervationally inaccessible observables (p.20), just as I do in Subsection 4.2 of Part III. When they come to the CTP formalism in Chapter 6, I see q-expectations and integrals over spacetime and frequency modes, but not over microscopic degrees of freedom.
 
  • Like
Likes dextercioby
  • #265
I have given my arguments why the very formalism of integrating over spatial (or temporal-spatial) volumes (four-volumes) is averaging/summing over many microscopic degrees of freedom. That's the view since the very beginning of statistical physics. In the field-theoretical context it's, as far as I know, first applied by Lorentz within his classical "electron theory". This is also what a macroscopic measurement device does, providing a single "pointer reading" in an experiment. I think that's very clear, and I don't know, how I can reformulate it to bring this point across better.

Obviously it's also hard to give an explanation, what your thermal interpretation is about in physical terms. The math is obviously the same as in standard quantum theory, based on the statistical interpretation. For me the only difference is that you forbid to call the formal procedure of taking the trace over observable operators times the statistical operator "averaging". What's lacking for me to understand this obviously crucial point is, where there is (a) the interpretation (i.e., where is the connection between the formalism and the physical meaning) and (b) in which sense I can understand it as "thermal" (which hinges for me also essentially on the statistical interpretation of the quantum state and the interpretation of the trace formula as averaging over many microscopic degrees of freedom).
 
  • #266
vanhees71 said:
I have given my arguments why the very formalism of integrating over spatial (or temporal-spatial) volumes (four-volumes) is averaging/summing over many microscopic degrees of freedom.
Yes, this is the well-known statistical interpretation, and I understand your arguments, so there is no point in repeating these.

But this interpretation is replaced in the thermal interpretation by alternative, philosophically less problematic imagery rooted in the formal mathematical core alone rather than in a reference to microscopic degrees of freedom.

In purely mathematical terms, spatiotemporal smearing is a filtering procedure that suppresses high frequency contributions to the field. This has an immediate and intuitive physical meaning since high frequency contributions in space or time cannot be resolved and thus must be suppressed somehow. One way of doing so is by regularization using spatial and temporal frequency cutoffs. Almost nobody introduces these as averaging, since the direct interpretation is much more intuitive.

General filtering is just a more adaptive linear regularization operation. The filtering procedure relevant for a localized measuring device (e.g., the human ear for measuring pressure oscillations) is a convolution with a function with negligible support outside the active region of the device, the Fourier transform of a function weighting each frequency according to its ability to cause a measurable response. The precise weight function to be used in the convolution is measurement dependent, and can be determined by calibration from measurements of prepared harmonic high frequency signals - audiometry in case of the ear, but nothing is special to the ear.

All this applies either on the operator level or on the level of q-expectations. The thermal interpretation treats the q-expectation ##\rho(x)=\langle j_0(x)\rangle## as an unregularized observable distribution-valued density field, and determines the regularized response of a particular measurement device by calibration, as in the case of the ear. All physical contents can be extracted in this way.

Nothing at all depends on the internal structure of the q-expectations. Their statistical interpretation as means over unmeasured microscopic degrees of freedom can simply be dispensed with, without any loss in understanding. Mathematically (i.e., with more than metaphorical invocation), this statistical interpretation works well anyway only in very simple situations such as very low density gases. Thus, like the particle picture, it is best regarded as an elementary precursor simplifying the real thing.

vanhees71 said:
Obviously it's also hard to give an explanation, what your thermal interpretation is about in physical terms. The math is obviously the same as in standard quantum theory, based on the statistical interpretation.
My interpretation is not at all hard. The math of standard quantum theory is completely independent of the interpretation, not anything special to the statistical interpretation. (It is shared by the Copenhagen interpretation, the many-histories interpretation, and all other fancy variations.)

vanhees71 said:
For me the only difference is that you forbid to call the formal procedure of taking the trace over observable operators times the statistical operator "averaging". What's lacking for me to understand this obviously crucial point is, where there is (a) the interpretation (i.e., where is the connection between the formalism and the physical meaning)
I gave in the last few mails very precise connections between the formalism and the physical meaning, sufficient to interpret everything done in the textbooks and in actual experiments in intuitive terms. I have no idea what you could possibly find missing.

Unless you insist on the identity ''physical = averaging over microscopic degrees of freedom''. But this is just your statistical interpretation of ''physical'', not the God-given meaning of the term!

vanhees71 said:
(b) in which sense I can understand it as "thermal" (which hinges for me also essentially on the statistical interpretation of the quantum state and the interpretation of the trace formula as averaging over many microscopic degrees of freedom).
I had explained ''thermal'' in Section 1 of Part II as follows:
Part II said:
Essential use is made of the fact that everything physicists measure is measured in a thermal environment for which statistical thermodynamics is relevant. This is reflected in the characterizing adjective ''thermal'' for the interpretation.
Specifically, the local values of an extensive field in local equilibrium thermal statistical mechanics (which describes all measurement devices) are q-expectation of the corresponding field operator, in a state that depends on the local temperature. Suitably filtered values (in the sense above) are directly measurable.
Invoking statistical concepts (which unfortunately still figure in the historical names of the concepts) do not add anything that would help in motivating or structuring the mathematical tools or their application to real experiments.

Thus the thermal setting provides a well-grounded point of departure for an interpretation, the thermal interpretation, in the same way as the double-slit experiment and the Stern-Gerlach experiment for few-particle experiments provided in the past a well-grounded point of departure for Born's statistical interpretation.
 
Last edited:
  • Like
Likes dextercioby
  • #267
So you say yourself what I say from the very beginning:

"Essential use is made of the fact that everything physicists measure is measured in a thermal environment for which statistical thermodynamics is relevant. This is reflected in the characterizing adjective ''thermal'' for the interpretation."

Also renormalization, which you quote above, interpreted in the Wilsonian sense, and that's the sense of the abstract mathematical procedure in physical terms, is precisely about coarse graining through averaging over microscopic degrees of freedom.

It's not as easy to get rid of statistics in the natural sciences. At least you cannot just use the very arguments of statistical thermodynamics and then forbid to use the usual probabilistic language just to hide the statistical foundations. It's a contradictio in adjecto!
 
  • #268
Thermal Interpretation Summary Redux
  1. Q-Expectations and Q-correlators are physical properties of quantum systems, not predicted averages. This makes these objects highly "rich" in terms of properties for ##\langle\phi(t_1)\phi(t_2)\rangle## is not merely a statistic for the field value, but actually a property itself and so on for higher correlators.
  2. To emphasize the point above remember that Probability theory is merely a formal construct whose elements are interpreted based on the context in which they are used. Thus we are not required to view things like ##Tr(\rho A)## as statistical statements fundamentally. Here in the Thermal Interpretation we view them as property assignment. They are merely propeties whose values have the same relation to each other as the formal relations of Probability theory.
  3. This richness of physical properties is not compatible with the notion of a system being purely decomposable into its subsystems in all cases. There are often properties such as a correlators like ##\langle\phi(t_1)\phi(t_2)\rangle## that are properties of the total system that don't arise from properties of subsystems.
  4. Since some properties are assigned to the system as a whole, which can be quite extended, they provide the nonlocal beables required by Bell's theorem. This is a combination of points above. Consider an extended two photon system. This has correlator properties like ##\langle AB\rangle## that are assigned to the whole system, no matter how extended it is and by the above these properties are not merely a property or combination of properties of any of the subsystems.

    Note in the Thermal Interpretation, once we have these nonlocal properties, violation of the CHSH inequalities for example has a simple explanation. There are simply four nonlocal total system properties:
    $$\langle AB\rangle, \langle BC\rangle, \langle CD\rangle, \langle AD\rangle$$
    for whom the following sum:
    $$\langle AB\rangle + \langle BC\rangle + \langle CD\rangle - \langle AD\rangle$$
    is greater than ##2##.
  5. Properties in the Thermal Interpretation are intrinsically "Fuzzy". For example for a particle its position property ##\langle q\rangle## has an associated property ##\Delta q##, in orthodox terminology called the uncertainty, that indicates the delocalised nature of the particle's position. Rather than the particle being located at points along a world line ##q(t)## it in fact constitutes a world tube centered on ##\langle q\rangle_t## and with width ##\Delta q_t##. (Question for @A. Neumaier , do higher statistics like skew and kurtosis give one further structure of the world tube?)

    All properties are blurred/fuzzy like this, not just position. Note in particular that spin for example is not discrete at the fundamental ontological level since ##\langle S_z\rangle## may take a continuous range of values. And thus we have a "spin tube" of width ##\Delta S_z##.

    For this reason there is a more basic notion of uncertainty to quantum properties akin to asking "What is the position of Paris". Since Paris is an extended object there is no precise answer to this question.
  6. Nonetheless as per standard scientific practice even this basic notion of uncertainty may be treated statistically.
  7. In measurements our devices (for reasons given in the next point) unfortunately only become correlated with a single point within the world tube or blurred range of a property. This gives measurements on quantum systems discrete results that don't faithfully represent the "tubes". Thus we must reconstruct them from multiple observations.
  8. The origin of this discreteness is in the metastability of the system-device-environment interaction. The device as a physical system can be partitioned into slow large scale and fast small scale modes. The space of slow large scale modes has the form of a set of disconnected manifolds. After interaction with the system the total state of the device initially faithfully records the uncertain quantity ##\langle A\rangle## however it is metastable. Noise from the environment causes it to quickly decay into one of the slow mode manifolds giving a discrete outcome not fully reflective of ##\langle A\rangle##.
    From our perspective, ignorant of the environmental dynamics, this constitutes a stochastically driven discrete outcome of a measurement.
 
  • Like
  • Informative
Likes vis_insita, Jimster41, Auto-Didact and 9 others
  • #269
vanhees71 said:
So you say yourself what I say from the very beginning:

"Essential use is made of the fact that everything physicists measure is measured in a thermal environment for which statistical thermodynamics is relevant. This is reflected in the characterizing adjective ''thermal'' for the interpretation."

Also renormalization, which you quote above, interpreted in the Wilsonian sense, and that's the sense of the abstract mathematical procedure in physical terms, is precisely about coarse graining through averaging over microscopic degrees of freedom.

It's not as easy to get rid of statistics in the natural sciences. At least you cannot just use the very arguments of statistical thermodynamics and then forbid to use the usual probabilistic language just to hide the statistical foundations. It's a contradictio in adjecto!
We talk of a vector space even if its elements are not vectors in the sense of little arrows or n-tuples. Similarly I refer to statistical thermodynamics for the results of the traditional theory without interpreting the items used there as mean - i.e., statistical mechanics as treated in Part II of my online book. You are allowed to use the old interpretation but then you are in the statistical interpretation and not in the thermal interpretation.
 
  • Like
Likes Auto-Didact
  • #270
Well, you always tell me how you don't interpret the "symbolism of atomic measurements" (Schwinger). Paradoxically, although claiming to provide a new interpretation, you never tell us its content. It's not enough to write down the formalism and forbid to use the 93 years old accepted probabilistic interpretation without providing a new one. Is @DarmMM 's posting #268 provides the correct interpretation of your interpretation, it's the more confirmed that it is still the usual probabilistic one too.
 
  • #271
vanhees71 said:
Paradoxically, although claiming to provide a new interpretation, you never tell us its content. It's not enough to write down the formalism and forbid to use the 93 years old accepted probabilistic interpretation without providing a new one. If @DarMM 's posting #268 provides the correct interpretation of your interpretation, it's the more confirmed that it is still the usual probabilistic one too.
You never listened to the content descriptions I gave (last in #263 and #266).

Of course I don't forbid anything; so you can always add probabilistic interpretations. DarMM's summary is essentially correct; I'll point to small corrections later. It never assumes anything probabilistic; thus one doesn't need it.

Unlike you, I allow people to dismiss statistical connotations in all situations where no actual measurement results are averaged over - without loss of physics but with a resulting improvement of the foundations. No irreducible probability, no dependence of the foundations on measurement, the same clarity in the association of predictions and experimental results. But the traditional foundational problems are gone!

You, of course, never saw any foundational problems in the statistical interpretation. This is because you are far more liberal in the use of intuition than strict adherence to the axioms in any statistical foundations would allow. At this level of liberality in the discussion we agree. But many others (e.g., @Stevendaryl, or Steven Weinberg) were never satisfied with such liberal (and hence poorly defined) foundations. The thermal interpretation is for those.

Our whole discussion shows that I didn't miss anything in the thermal interpretation, and really covered the actual physical usage of the quantum formalism.
 
Last edited:
  • Like
Likes Jimster41
  • #272
DarMM said:
Thermal Interpretation Summary Redux
Excellent summary, thanks!
DarMM said:
7. In measurements our devices (for reasons given in the next point) unfortunately only become correlated with a single point within the world tube or blurred range of a property. This gives measurements on quantum systems discrete results that don't faithfully represent the "tubes". Thus we must reconstruct them from multiple observations.
The ''single point'' here is of course also a blurred point only, as a detector always responds to a small space-time region and not to a point only. The measurement result will often be discrete, however, due to the reasons given in point 8.

It is like measuring the rate of water flow into a basin by the number of buckets per unit time needed to keep the water at a roughly fixed level of height. As long as there is enough flow the bucket is very busy and the flow is measured fairly accurately. But at very low rates it is enough to occasionally take out one bucket full of water and the bucket number is a poor approximation of the flow rate unless one takes very long times.

Photocells and Geiger counters act as quantum buckets, by the same principles.
 
Last edited:
  • Like
Likes dextercioby and DarMM
  • #273
DarMM said:
Question for @A. Neumaier , do higher statistics like skew and kurtosis give one further structure of the world tube?
The world tube is just a visualization, even a fuzzy one since it cannot be said how many sigmas the world tube is wide. I don't think skewness and kurtosis are valuable on this fuzzy visualization level.

Rather, the next level of naturally visualizable detail would be to think of the particle's world tube as being actually a moving cloud of mass (or energy, or charge, etc.), with mass density ##\rho(z)=\langle m\delta(x_k-z)\rangle## for a distinguishable particle ##k## inside a multiparticle system. For a single particle state, this would be ##\rho(z)=m|\psi(z)|^2##, recovering Schrödinger's intuition about the meaning of the wave function. Of course, this is already the field picture, just in first quantized form.

Mathematically, of course, there is no world tube at all but just a host of loosely related beables. One family of them is the reference path ##\langle x(t)\rangle##. It has a clear intuitive meaning (as the path one sees at low resolution), just as the reference path of a classical donut traced out by the center of mass, though in both cases nothing is precisely at the reference path.
 
  • Like
Likes DarMM
  • #274
A. Neumaier said:
For a single particle state, this would be ρ(z)=m|ψ(z)|2ρ(z)=m|ψ(z)|2\rho(z)=m|\psi(z)|^2, recovering Schrödinger's intuition about the meaning of the wave function. Of course, this is already the field picture, just in first quantized form

I most definitely agree with this picture based on the example I gave in post #255. The only thing is that your system does not tell exactly what is going on in detail in that small region, my idea supplements and give the interpretation of the "line" picking distances as energy which upon calculation adds to mass according to good old Einstein's formula. :partytime:

P.S. I think you were right Arnold. The wavefunction is nothing, only density has a physical reality.
 
Last edited:
  • #275
ftr said:
does not tell exactly what is going on in detail in that small region
Nothing exact can be said. Differences of the order of the uncertainty are physically meaningless.
 
  • #278
A. Neumaier said:
Unlike you, I allow people to dismiss statistical connotations in all situations where no actual measurement results are averaged over - without loss of physics but with a resulting improvement of the foundations. No irreducible probability, no dependence of the foundations on measurement, the same clarity in the association of predictions and experimental results. But the traditional foundational problems are gone!
Again, just to understand, because I think this is the main source of our mutual misunderstanding:

Let's take my example for the measurement of a current with a galvanometer. From a quantum-theoretical point of view electric current consists of the motion of many conduction electrons in a wire at room temperature.

My point of view is that at a microscopic level you have a many-electron system, and the current density is described by the operator
$$\hat{j}^{\mu}(x)=-e \bar{\psi}(x) \gamma^{\mu} \psi(x).$$
What's measured by the galvanometer is the expectation value
$$J^{\mu}(x) = \mathrm{Tr} (\hat{\rho} \hat{j}(x)),$$
where ##\hat{\rho}## is a statistical operator at (or close) to thermal equilibrium, i.e., something like ##\propto \exp[-(\hat{H}-\mu \hat{Q})/T]##.
Now the galvanometer provides the current,
$$I=\int \mathrm{d} \vec{f} \cdot \vec{J}$$
right away as a pointer reading, because it is providing the above formally noted expectation value due to precisely resolving the measured quantity at this level of resolution.

This is within the standard statistical interpretation of the formalism. Now, how does your thermal interpretation explain this situation, if it denies the validity of the statistical arguments.
 
  • #279
vanhees71 said:
Again, just to understand, because I think this is the main source of our mutual misunderstanding:

Let's take my example for the measurement of a current with a galvanometer. From a quantum-theoretical point of view electric current consists of the motion of many conduction electrons in a wire at room temperature.

My point of view is that at a microscopic level you have a many-electron system, and the current density is described by the operator
$$\hat{j}^{\mu}(x)=-e \bar{\psi}(x) \gamma^{\mu} \psi(x).$$
What's measured by the galvanometer is the expectation value
$$J^{\mu}(x) = \mathrm{Tr} (\hat{\rho} \hat{j}(x)),$$
where ##\hat{\rho}## is a statistical operator at (or close) to thermal equilibrium, i.e., something like ##\propto \exp[-(\hat{H}-\mu \hat{Q})/T]##.
Now the galvanometer provides the current,
$$I=\int \mathrm{d} \vec{f} \cdot \vec{J}$$
right away as a pointer reading, because it is providing the above formally noted expectation value due to precisely resolving the measured quantity at this level of resolution.

This is within the standard statistical interpretation of the formalism. Now, how does your thermal interpretation explain this situation, if it denies the validity of the statistical arguments.
In the thermal interpretation ##J^{\mu}(x)## (or a smeared version thereof) is a quantity itself not the average outcome for an indeterminate quantity.
 
  • #280
Hm, well, that's just the usual sloppy language of some textbooks on QM to identify the expectation values with the actually observed quantities. I think there's much more behind Arnold's interpretation, which however I still miss to understand :-(.
 
  • #281
The Thermal Interpretation isn't identifying ##J^{\mu}(x)## with the observed quantity, it's saying that it is an actual property of the system.

The whole notion of statistics in the TI arises because, as mentioned in my summary, the observed quantities don't accurately reflect this underlying ontic quantity.
 
  • #282
vanhees71 said:
how does your thermal interpretation explain this situation
OK. Let us reconsider your example for the measurement of a current with a galvanometer. From a quantum field theoretical point of view, the electric current consists of the motion of the electron field in a wire at room temperature. The thermal interpretation says that at any description level you have an electron field, and the theoretically exact current density is described by the distribution-valued beable (q-expectation)
$$J^{\mu}(x) = \mathrm{Tr} (\hat{\rho} \hat{j}(x))$$
determined by the current operator
$$\hat{j}^{\mu}(x)=-e :\bar{\psi}(x) \gamma^{\mu} \psi(x):,$$
where the colons denote normal ordering and ##\hat{\rho}=e^{-S/k_B}## is the density operator describing the exact state of the galvanometer plus wire. (I usually don't write hats over the operators, but kept them when copying your formulas.)

At this level of description there is no approximation at all; the latter is introduced only when one replaces the exact ##S## by a numerically tractable approximation. At or close to thermal equilibrium, it is well-established empirical knowledge that we have ##S \approx (H+PV-\mu N)/T## (equality defines exact equilibrium). We can substitute this (or a more accurate nonequilibrium) approximation into the defining formula for ##J(x)## to compute a numerical approximation.

Measurable by the galvanometer is a smeared version
$$J^{\mu}_m(x) =\int dz\ h(z) J^{\mu}(x+z)$$
of the exact current density, where ##h(z)## is a function that is negligible for ##z## larger than the size of the current-sensitive part of the galvanometer. The precise ##h## can be found by calibration (as discussed earlier). You didn't explain what your ##\vec f## refers to, but taking your formula as being correct with an appropriate ##\vec f##, the galvanometer provides (ignoring the reading uncertainty) the current
$$I=\int \mathrm{d} \vec{f} \cdot \vec{J_m}$$
right away as a pointer reading.

This is without the standard statistical interpretation of the formalism. I made nowhere use of any statistical argument, treating the trace simply as a calculational device - as you do it for defining field correlations.

The only averaging is the smearing needed for mathematical reasons to turn the distribution-valued current into an observable vector, and for physical reasons since the galvanometer is insensitive to very high spatial or temporal frequencies. Note that this smearing has nothing to do with coarse-graining or averaging over microscopic degrees of freedom: It is also needed in (already coarse-grained) classical field theories. For example, in hydromechanics, the Navier-Stokes equations generally have only weak (distributional) solutions that make numerical sense only after smearing. Thus the quantum situation is in the thermal interpretation no different from the classical situation (except that there are many more beables than in the classical case).

vanhees71 said:
Hm, well, that's just the usual sloppy language of some textbooks on QM to identify the expectation values with the actually observed quantities. I think there's much more behind Arnold's interpretation, which however I still miss to understand :-(.
I can see nothing sloppy in my arguments.
 
Last edited:
  • Like
Likes Mentz114, DarMM and dextercioby
  • #283
My problem is that there is no physical motivation given for all the formal mathematical manipulations you do. It's of course all right formally, but to understand, why you are doing all these manipulations you need an interpretation, which you still don't provide since you simply say it's not the usual statistical interpretation behind it, but you still don't say, what instead is behind these manipulations. I don't think we come to a conclusion about these different pictures of what an interpretation in principle should provide, repeating the same arguments over and over again.
 
  • #284
stevendaryl said:
I consider orthodox quantum mechanics to be incomplete (or maybe inconsistent) [...]
Alice and Bob prepare an electron so that it is a superposition of spin-up and spin-down. Alice measures the spin. According to orthodox quantum mechanics, she either gets spin-up with such-and-such probability, or she gets spin-down with such-and-such probability.
But Bob, considering the composite system of Alice + electron
Orthodox quantum mechanics applies without problems when only a single quantum system is considered, as the Copenhagen interpretation requires.

Inconsistencies appear (only) when one considers simultaneously a quantum system and a subsystem of it (as in your example). The standard foundations (which work with pure states only and treat mixed states as states of incomplete knowledge, with classical uncertainty about the true pure state) simply do not cater appropriately for this situation. The reason is discussed in detail in Subsection 3.5 of Part I of my series of papers. In short, the standard foundations assert

(S1) The pure state of a system (at a given time) encodes everything that can be said (or ''can be known'') about the system at this time, including the possible predictions at later times, and nothing else.

On the other hand, common sense tells us

(S2) Every property of a subsystem is also a property of the whole system.

Now (S1) and (S2) imply

(S3) The pure state of a system determines the pure state of all its subsystems.

But (S3) is denied by the standard foundations, since knowing the pure state of an entangled system does not tell us anything at all about the pure state of its constituents. Nothing in the literature tells us anything about how to determine the pure state of a subsystem from the pure state of a system (except in the idealized, unentangled situation). Not even that - not a single property of a subsystem alone can be deduced from the pure state of the full system if the latter is completely entangled (which is the rule for multiparticle systems)!

The conclusion is that in the standard foundations, we have in place of (S2)

(S??) No property of a subsystem is also a property of the whole system.

This absurd conclusion causes the well-known foundational problems.

Not only that - taken at face value, (S??) would mean that the quantum state of a physics lab has nothing to do with the quantum states of the equipment in it, and of the particles probed there!

This is very strange for a science such as physics that studies large systems primarily by decomposing them into its simple constituents and infers properties of the former from collective properties of the latter.

The thermal interpretation avoids all these problems by taking the state to be an arbitrary monotone linear functional defining q-expectations (thus allowing for mixed states), and the (functions of ) q-expectations of operators as the properties of a system. The properties of a subsystem are then simply the (functions of) q-expectations of operators that act on the subsystem alone. Thus the properties (S1)-(S3) - with the word ''pure'' dropped - are trivially valid!
 
Last edited:
  • Like
Likes Mentz114 and dextercioby
  • #285
A. Neumaier said:
Orthodox quantum mechanics applies without problems when only a single system is considered, as the Copenhagen interpretation requires.

I would say, on the contrary, that the Copenhagen interpretation always involves at least two systems: the system being measured and the measuring device.

The thermal interpretation avoids all these problems by taking the state to be an arbitrary monotone linear functional (allowing for mixed states), and the (functions of ) q-expectations of operators as the properties of a system. The properties of a subsystem are then simply the (functions of) q-expectations of operators that act on the subsystem alone. Thus the properties (S1)-(S3) - with the word ''pure'' dropped - are trivially valid!

Yes, I appreciate that. But I don't completely understand it. If you have, for example, an entangled pair of spin-1/2 particles, ##\sqrt{\frac{1}{2}} (|ud\rangle - |du\rangle)##, and someone measures the spin of the first particle, then the result would seem to be something that is not computable from expectations and correlations.
 
  • #286
vanhees71 said:
My problem is that there is no physical motivation given for all the formal mathematical manipulations you do. It's of course all right formally, but to understand, why you are doing all these manipulations you need an interpretation, which you still don't provide since you simply say it's not the usual statistical interpretation behind it, but you still don't say, what instead is behind these manipulations.
You have the same problem when introducing in your research work Green's functions as purely calculational tools, deferring the interpretation in terms of experiments to results derived from the formal manipulations.

vanhees71 said:
I don't think we come to a conclusion about these different pictures of what an interpretation in principle should provide, repeating the same arguments over and over again.
I demand from an interpretation only that it gives a clear interface between the theoretical predictions and the experimental record, and in this respect the thermal interpretations serves very well.

People were always doing those formal manipulations where history has shown that it works, while the intuition associated with the formal manipulations changed. Maxwell derived his equations from intuition about a mechanical ether; we think of this today as crutches that are no longer needed, and in fact are detrimental for a clear understanding.

Thus whether or not someone accompanies the formal manipulations with intuitive blabla - about microscopic degrees of freedom, or virtual particles popping in and out of existence, or Bohmian ghost variables, or many worlds, or whatever - is a matter of personal taste. I prefer to cut all this out using Ockham's razor, and leave it to the subjective sphere of individual physicists.
 
Last edited:
  • Like
Likes Mentz114 and dextercioby
  • #287
The orthodox quantum mechanics (which for me is the formalism + the minimal statistical interpretion, i.e., that flavors of Copenhagen that do not envoke some esoterical "quantum-classical cut" nor an even more esoterical "collapse of the state") work perfectly well, because orthodox quantum mechanics does not claim that there are only pure states.

Of course, you have maximal possible knowledge about a system only, if you have prepared it in a pure state. If you now prepare a composite system, described in a product space ##\mathcal{H}_1 \otimes \mathcal{H}_2##, with the Hilbert spaces referring to the parts, in a pure state ##\hat{\rho}=|\Psi \rangle \langle \Psi \rangle## you have maximal possible knowledge about this composite systems.

If you choose to ignore one of the parts, by definition you through away information, by choosing to ignore the other part of the system. Then, in general, of course the partial system you choose to consider, is not prepared in a pure state, but in the state given by the partial trace, i.e., if you only like to take notice of part 1 of the system, you describe it, according to the basic definitions of probability theory, by the mixed state given by the partial trace,
$$\hat{\rho}_1=\mathrm{Tr}_2 \hat{\rho}.$$
 
  • #288
@A. Neumaier what do you consider the open problems of the Thermal Interpretation?

Establishing that devices in certain canonical scenarios have the metasability properties and slow mode disconnected manifolds required? And also that these in fact lead to discrete outcomes whose long run frequencies have an average equal to the true value ##\langle A\rangle##? i.e. that in certain states the device will fall into one slow mode more frequently for example.
 
  • #289
A. Neumaier said:
You have the same problem when introducing in your research work Green's functions as purely calculational tools, deferring the interpretation in terms of experiments to results derived from the formal manipulations.I demand from an interpretation only that it gives a clear interface between the theoretical predictions and the experimental record, and in this respect the thermal interpretations serves very well.

People were always doing those formal manipulations where history has shown that it works, while the intuition associated with the formal manipulations changed. Maxwell derived his equations from intuition about a mechanical ether; we think of this today as crutches that are no longer needed, and in fact are detrimental for a clear understanding.

Thus whether or not someone accompanies the formal manipulations with intuitive blabla - about microscopic degrees of freedom, or virtual particles popping in and out of existence, or Bohmian ghost variables, or whatever - is a matter of personal taste. I prefer to cut all this out using Ockham's razor, and leave it to the subjective sphere of individual physicists.
I don't have a problem to define Green's functions. Why should I have a problem? They are calculational tools, as are, e.g., the potentials in electromagnetism which are not observable as well. There are plenty of calculational objects in theoretical physics which do not directly refer to observables. Why should I have problems introducing them to help to calculate predictions for observables?

It's also clear that you can get correct theories from wrong heuristics, as your example of the historical derivation of Maxwell's equations from mechanistic aether models show. Of course, here we have also a perfect example, how physicists are forced to change their heuristics and the interpretation with progress in science: Maxwell himself finally abandoned his own mechanistic aether interpretation. As is well known, the aether was completely abandoned only much later after Einstein's discovery of special relativity, which also lead to a change of interpretation in terms of relativistic space-time descriptions.

The only thing, I still find lacking with your "thermal interpretation" is that it doesn't provide an interpretation at all but just tries to sell the formalism as the interpretation. It precisely does not provide what you yourself defined as the purpose of any interpretation, namely the link between the formalism and the observable facts in the lab!
 
  • #290
stevendaryl said:
the Copenhagen interpretation always involves at least two systems: the system being measured and the measuring device.
But the measurement device is treated classically, not in terms of a quantum state.
stevendaryl said:
Yes, I appreciate that. But I don't completely understand it. If you have, for example, an entangled pair of spin-1/2 particles, ##\sqrt{\frac{1}{2}} (|ud\rangle - |du\rangle)##, and someone measures the spin of the first particle, then the result would seem to be something that is not computable from expectations and correlations.
The measurement result is primarily a property of the detector, and is computable - up to measurement errors - from the state of the detector (e,g., from the position ##\langle q\rangle## of the center of mass ##q## of the pointer).

As an approximate measurement of the first particle, it is a poor approximation (binary, with large uncertainty) of the the z-component of the spin of the first particle, which - in the thermal interpretation - is the real number ##\langle J_3\rangle=\langle i\sigma_3\rangle##.

To reduce the uncertainty one can measure the spin of a whole ensemble of ##N## particles and compute the ensemble average, thereby reducing the uncertainty by a factor of ##\sqrt{N}##. For large ##N## this provides a statistical interpretation of the q-expectation.
 
Last edited:
  • Like
Likes dextercioby
  • #291
A. Neumaier said:
But the measurement device is treated classically, not in terms of a quantum state.

Okay. But that has to be an approximation, because the measurement device is made out of electrons and protons and photons, just like the system being measured.

The measurement result is primarily a property of the detector, and is computable - up to measurement errors - from the state of the detector (e,g., from the position ##\langle q\rangle## of the center of mass ##q## of the pointer).

I don't see how that can be true. In an EPR twin-pair experiment, if Alice measures spin-up for her particle along the z-axis, then Bob will measure spin-down along that axis. The details of Bob's measuring device seem completely irrelevant (other than the fact that it actually measures the z-component of the spin).
 
  • #292
It's of course not true. A measurement device by definition and construction (that's half of the very art of experimentalists, the second being the art of evaluating the measurement results properly, including an estimate of the statistical (sic!) and systematic errors) after interacting with the system to be measured, provides some information of the measured observable of the system.

The most simple example is the Stern-Gerlach experiment. The measurement device is the magnet, leading to an (approximately!) entanglement of the position of the particle (pointer variable) and the spin component in the direction of the (homogeneous part of) the magnetic field. This means to measure that spin component you just have to collect the particles (in the original Frankfurt experiment Ag atoms) on a screen and look at the position. As you see, here the pointer variable is in the very particle itself, namely its position. There's no need for any classical approximation whatsoever (except for the final step of "collecting" the particle on the screen).
 
  • #293
DarMM said:
what do you consider the open problems of the Thermal Interpretation?

Establishing that devices in certain canonical scenarios have the metasability properties and slow mode disconnected manifolds required? And also that these in fact lead to discrete outcomes whose long run frequencies have an average equal to the true value ##\langle A\rangle##? i.e. that in certain states the device will fall into one slow mode more frequently for example.
Yes, this kind of things, which are argued in my papers only qualitatively. I listed some open problems at the end of the Conclusions of Part III. Also, the AB&N work is too complex to be studied in depth by many, and finding simpler scenarios that can be analyzed in a few pages (so that they are suitable for courses in quantum mechanics) would be worthwhile.
 
  • #294
vanhees71 said:
It's of course not true. A measurement device by definition and construction (that's half of the very art of experimentalists, the second being the art of evaluating the measurement results properly, including an estimate of the statistical (sic!) and systematic errors) after interacting with the system to be measured, provides some information of the measured observable of the system.

The most simple example is the Stern-Gerlach experiment. The measurement device is the magnet, leading to an (approximately!) entanglement of the position of the particle (pointer variable) and the spin component in the direction of the (homogeneous part of) the magnetic field. This means to measure that spin component you just have to collect the particles (in the original Frankfurt experiment Ag atoms) on a screen and look at the position. As you see, here the pointer variable is in the very particle itself, namely its position. There's no need for any classical approximation whatsoever (except for the final step of "collecting" the particle on the screen).

I would say that there SHOULD be no need for any classical approximation, if quantum mechanics were consistent and complete.
 
  • #295
It's of course a challenge, but I don't think that there are any principle problems in understanding macroscopic measurement devices (as in this example the screen used to collect the Ag atoms) quantum theoretical since measurement devices are just macroscopic physical systems as anything around us, and these are quite well understood in terms of quantum many-body theory. I'd say condensed-matter physics is one of the most successful branches of physics, and its very successful research is precisely dealing with the quantum theoretical understanding of the classical properties of macroscopic systems, and measurement devices are just such macroscopic systems manipulated by experimentalists and engineers to enable their use as measurement devices.

That's why I was so inclined to the "thermal interpretation" at first, because I thought it just takes this usual point of view of the macroscopic behavior being described by quantum-many body theory in terms of appropriate coarse-graining procedures, which is a generically statistical-physics argument. Then I learned in this discussion, that I completely misunderstood the three papers, because it's precisely how I should NOT understand the interprtation, but now I'm lost since I've no clue what the interpretation now is meant to be.
 
  • #296
stevendaryl said:
Okay. But that has to be an approximation, because the measurement device is made out of electrons and protons and photons, just like the system being measured.
Yes, but Copenhagen says nothing about the nature of this approximation. It takes definite classical measurement values as an irreducible fact. The measurement problem begins precisely when one wants to derive this fact rather than postulate it.

stevendaryl said:
I don't see how that can be true. In an EPR twin-pair experiment, if Alice measures spin-up for her particle along the z-axis, then Bob will measure spin-down along that axis. The details of Bob's measuring device seem completely irrelevant (other than the fact that it actually measures the z-component of the spin).
Well, you surely agree that the pointer position can be calculated (in principle) from the state of the measurement device at the time the reading is done. This has nothing at all to do with what is measured, it is only the fact that the pointer is somewhere where it can be read.

What you and I don't see is how precisely Bob's detector knows which pointer reading to display to conform to the correlated quantum statistics. Somehow, Nature does it, and since we haven't found any deviation from quantum physics, it is presumably a consequence of the Schrödinger dynamics of the state of the universe. In Subsections 4.4 and 4.5 of Part II, I discussed the sense in which this can be partially understood: the extendedness of quantum systems (entangled long-distance states are very fragile extended systems) together with conditional knowledge make things reasonably intelligible to me. In any case, it is in full agreement with the requirements of relativity theory.
 
Last edited:
  • #297
stevendaryl said:
I would say that there SHOULD be no need for any classical approximation, if quantum mechanics were consistent and complete.

The basic postulate of the minimal interpretation of quantum mechanics is: If you measure an observable, then you will get an eigenvalue of the corresponding operator. But if measurement is definable in terms of more basic concepts, then that rule should be expressible in terms of those more basic concepts.

So for example, an observable is measured when it interacts with a macroscopic system (the measuring device) so that the value of that observable can be "read off" from the state of the macroscopic system. In other words, the microscopic variable is "amplified" to make it have a macroscopic effect.

So the rule about measurement producing an eigenvalue boils down to a claim that a macroscopic measurement device is always in one of a number of macroscopically distinguishable states.

So the minimal interpretation of QM, it seems to me, is equivalent to the following:

Any system, no matter how complex, evolves according to unitary evolution with the exception that a macroscopic system is always in one of a number of macroscopically distinguishable states. If unitary evolution would put the system into a superposition of macroscopically distinguishable states, then the system will pick one, with probability given by the square of the amplitude corresponding to that possibility.

This way of stating the Born rule is, as far as I know, exactly equivalent in its empirical content to the minimal interpretation. But it differs in that it does not claim that measurements produce eigenvalues. That is a derivable consequence (from the definition of "measurement" plus the rules for unitary evolution). So I actually think that this formulation is better. But it explicitly makes a distinction between macroscopic and microscopic observables. This distinction is already implicit in the minimal interpretation, but is hidden by the lack of a clear definition of "measurement".
 
  • #298
vanhees71 said:
I don't have a problem to define Green's functions. Why should I have a problem? They are calculational tools, as are, e.g., the potentials in electromagnetism which are not observable as well. There are plenty of calculational objects in theoretical physics which do not directly refer to observables. Why should I have problems introducing them to help to calculate predictions for observables?
But why do you demand more for my calculational tools, the q-expectations? Justify them exactly in the same way as you justify your definition of Green's function, and you'll understand me.

vanhees71 said:
The only thing, I still find lacking with your "thermal interpretation" is that it doesn't provide an interpretation at all but just tries to sell the formalism as the interpretation. It precisely does not provide what you yourself defined as the purpose of any interpretation, namely the link between the formalism and the observable facts in the lab!
I only claim that some of the q-expectations give predictions for actual observations, and how these are to be interpreted is fully specified in my posts #263, #266, and #278 (and elsewhere).
 
Last edited:
  • #299
A. Neumaier said:
Well, you surely agree that the pointer position can be calculated (in principle) from the state of the measurement device at the time the reading is done. This has nothing at all to do with what is measured, it is only the fact that the pointer is somewhere where it can be read.

Let's look at the specific case: You have an electron sent through a Stern-Gerlach device. Depending on the electron's spin, it either is deflected left, and makes a black dot on a photographic plate on the left, or is deflected right, and makes a black dot on a photographic plate on the right.

If the electron is deflected left, then it WILL make a dot on the left plate. So the details of the photographic plate just don't seem relevant, other than the fact that its atoms are in an unstable equilibrium so that a tiny electron can trigger a macroscopic state change.
 
  • #300
stevendaryl said:
Let's look at the specific case: You have an electron sent through a Stern-Gerlach device. Depending on the electron's spin, it either is deflected left, and makes a black dot on a photographic plate on the left, or is deflected right, and makes a black dot on a photographic plate on the right.

If the electron is deflected left, then it WILL make a dot on the left plate. So the details of the photographic plate just don't seem relevant, other than the fact that its atoms are in an unstable equilibrium so that a tiny electron can trigger a macroscopic state change.
I completely agree. But this doesn't invalidate what I said:

That there is a dot on the left, say, is a property of the plate that is computable from the state of the plate at the time of looking at it to record the event. For the state of the plate with a dot on the right is quite different from the state of the plate with a dot on the left or that without a dot. Hence the state differentiates between these possibilities, and hence allows one to determine which possibility happened.

The only slighly mysterious thing is why Alice can predict Bob's measurement. Here I don't have a full explanation, but only arguments that show that nothing goes wrong with relativistic causality.

Note also that the prediction comes out correct only when the entangled state is undisturbed and the detector is not switched off at the time the photon hits - things that Alice cannot check until having compared notes with Bob. Thus her prediction is not a certainty but a conditional prediction only.
 
  • Like
Likes Jimster41

Similar threads

Back
Top