Quantum Bayesian Interpretation of QM

Salman2
Messages
94
Reaction score
0
Any comments (pro-con) on this Quantum Bayesian interpretation of QM by Fuchs & Schack ?:

http://arxiv.org/pdf/1301.3274.pdf
 
Physics news on Phys.org
Yes, I'd also like to know if anyone has any insights on this new model, dubbed the "QBism" model. The general idea is that the quantum wave function does NOT represent any actuality in the real physical world. It is an abstraction of the mind...and it goes from there. The arxiv article Salman2 posted is called the No nonsense version, which sounds like a good initial review. However, for those who want an even briefer survey should check the most recent SciAm issue:

http://www.scientificamerican.com/a...tum-beyesnism-fix-paradoxes-quantum-mechanics
 
"The general case of conscious perception is the negative perception, namely, 'perceiving the stone as not grey'...
"Consciousness is the subjective form involved in feeling the contrast between the 'theory' which may be erroneous and the fact which is given."

Alfred North Whitehead, Process and Reality, ISBN 0-02-934570-7, p. 161.

"In other words, consciousness enters into the subjective forms of feelings, when those feelings are components in an integral feeling whose datum is the contrast between a nexus which is, and a proposition which in its own nature negates the decision of its truth or falsehood."

p. 261

Again, H P Stapp sees Whitehead as providing a container for QM. I agree. If he had lived a little longer, he would have been in the QM Pantheon.

CW
 
Last edited:
Chris Fields offer a critique of the QBism model presented by Fuchs based on the nature of the wavefunction of the agent (observer):

http://arxiv.org/pdf/1108.2024v2.pdf

Here from the end of the paper is the major objection of Fields to QBism:

"QBism provides no physical distinction between observers and the systems they observe, treating all quantum systems as autonomous agents that respond to observations by updating beliefs and employ quantum mechanics as a “users’ manual” to guide behavior. However, it treats observation itself as a physical process in which an “observer” acts on a “system” with a POVM and the system” selects a POVM component as the “observer’s experience” in return. This requirement renders the assumption that systems be well-defined - i.e. have constant d-impossible to implement operationally. It similarly forces the consistent QBist to regard the environment as an effectively omniscient observer, threatening the fundamental assumption of subjective probabilities and forcing the conclusion that QBist observers cannot segment their environments into objectively separate systems."

==

Another paper by Fields, discussion of QBism starts on p. 27:

http://arxiv.org/pdf/1108.4865.pdf
 
Last edited:
Salman2 said:
Any comments (pro-con) on this Quantum Bayesian interpretation of QM by Fuchs & Schack ?:

http://arxiv.org/pdf/1301.3274.pdf

Epistemic-Epistemic view of the wave function.
 
Subjective Reality-

The discussion of QBism poses iepistemological, and semantic problems for the reader. The subtitle- It's All ln Your Mind- is a tautology. Any theory or interpretation of observed physical phenomena is in the mind, a product of the imagination, or logical deduction, or some other mental process. Heisenberg ( The Physical Principles of the Quantum Theory), [n discussing the uncertainty principle, cautioned that human language permits the construction of sentences that have no content since they imply no experimentally observable consequences , even though they may conjure up a mental picture. He particularly cautioned against the use of the term, " real " in relation to such statements. as is done in the article. Mr. von Burgers also described QBism as representing subjective beliefs-whose? Bertrand Russell (Human Knowlrdge, Its Scope and Limits) described " belief" as a word not easy to define. It is certainly not defined in the context of the article.
,
Heisenberg also showed that the uncertainty principle and several other results of quantum mechanical theory could be deduced without reference to a wave function, so this aspect of the new interpretations is not unique. Similarly Feynman (QED, The Strange Theory of Light and Matter) dealt with the diffraction of light through a pair of slots, by a formulation based on the actions of photons without reference to wave functions. The statement in the article that the wave function is
" only a tool" to enable mathematical calculations is puzzling- any theoretical formulation of quantum mechanics is a tool for mathematical calculations relating to the properties of physical systems.

In spite of the tendency in Mr. von Burgers' article to overplay the virtues of QBism relative to other formulations, as an additional way to contemplate quantum mechanics, it has potential value, As Feynman ( The Character of Physical Law) stated, any good theoretical physicist knows six or seven theoretical representations for exactly the same physics. One or another of these may be the most advantageous way of contemplating how to extend the theory into new domains and discover new laws. Time will tell.

Alexander
 
Last edited by a moderator:
wayseeker said:
The discussion of QBism poses iepistemological, and semantic problems for the reader. ...

Welcome to PhysicsForums, Alexander!

Are you familiar with the PBR theorem? Although I can't say I fully understand the examples in the OP's QBism paper, it seems to flow directly opposite to PBR. One says the wave function maps directly to reality, the other says it does not.
 
Some interesting remarks on Bayes' Theorem

http://www.stat.columbia.edu/~gelman/research/published/badbayesmain.pdf
"Bayesian inference is one of the more controversial approaches to statistics, with both the promise and limitations of being a closed system of logic. There is an extensive literature, which sometimes seems to overwhelm that of Bayesian inference itself, on the advantages and disadvantages of Bayesian approaches"---------
"Bayes' Theorem is a simple formula that relates the probabilities of two different events that are conditional upon each other"

sound familiar, no ?
(in physics, i mean)
 
Last edited:
DrChinese said:
Are you familiar with the PBR theorem? Although I can't say I fully understand the examples in the OP's QBism paper, it seems to flow directly opposite to PBR. One says the wave function maps directly to reality, the other says it does not.

I have gone through the PBR theorem and my view is exactly the same as Matt Leifer:
http://mattleifer.info/2011/11/20/can-the-quantum-state-be-interpreted-statistically/

He divides interpretations into three types:

1. Wavefunctions are epistemic and there is some underlying ontic state. Quantum mechanics is the statistical theory of these ontic states in analogy with Liouville mechanics.
2. Wavefunctions are epistemic, but there is no deeper underlying reality.
3. Wavefunctions are ontic (there may also be additional ontic degrees of freedom, which is an important distinction but not relevant to the present discussion).

PBR says nothing about type 2 - in fact the paper specifically excludes it. What it is concerned about is theories of type 1 and 3 - it basically says type 1 is untenable - its really type 3 in disguise.

That's an interesting result but I am scratching my head why its considered that important. Most interpretations are type 2 (eg Copenhagen and the Ensemble interpretation), many others are type 3 (eg MWI and BM) and only a few type 1.

Maybe I am missing something but from what I can see its not that big a deal.

Thanks
Bill
 
Last edited:
  • #10
Regarding the Quantum Baysean interpretation its a perfectly good way of coming to grips with the probability part of QM.

Normally that is done by means of an ensemble view of probability which talks about the proportion of a very large number of similar systems and the proportion with a particular property or whatever is the probability. This is a view very commonly used in applied math. But Baysian probability theory (perhaps framework is a better word) - is just as valid. In fact there is some evidence it leads to a slicker axiomatic formulation:
http://arxiv.org/pdf/quant-ph/0210017v1.pdf

Even if it isn't as slick I prefer the ensemble view because of its greater pictorial vividness.

IMHO its not really an issue to get too concerned about. In applying probability to all sorts of areas the intuitive view most have is from my experience more than adequate without being strict about it.

Thanks
Bill
 
Last edited:
  • #11
DiracPool said:
Yes, I'd also like to know if anyone has any insights on this new model, dubbed the "QBism" model. The general idea is that the quantum wave function does NOT represent any actuality in the real physical world.

As far as I can see its simply the ensemble interpretation in another guise - where the pictorial vividness of an ensemble is replaced by beliefs about information.

Information seems to be one of the buzz things in physics these days but personally I can't see the appeal - although am willing to be convinced.

You might be interested in the following where QM is derived from all systems with the same information carrying capacity are the same (plus a few other things):
http://arxiv.org/pdf/0911.0695v1.pdf

It's my favorite foundational basis of QM these days - but I have to say it leaves some cold.

The interesting thing though is if you remove information from the axioms and say - all systems that are observationally the same are equivalent it doesn't change anything in the derivation - which sort of makes you wonder.

Thanks
Bill
 
  • #12
Well, maybe I'm just too much biased by my training as a physicist to make sense of the whole Baysian interpretation of probabilities. In my opinion this has nothing to do with quantum theory but with any kind of probabilistic statement. It is also good to distinguish some simple categories of content of a physical theory.

A physical theory, if stated in a complete way like QT, is first some mathematical "game of our minds". There is a well-defined set of axioms or postulates which give a formal set of rules which establishes how to calculate abstract things. In QT that's the "state of the system", given by a self-adjoint positive semidefinite trace-class operator on a (rigged) Hilbert space, an "algebra of observables", represented by self-adjoint operators, and a Hamiltonian among the observables that defines the dynamics of the system. That's just the formal rules of the game. It's just a mathematical universe, you can make statements (prove theorems), do calculations. I think this part is totally free of interpretational issues, because no connection to the "real world" (understood as reproducible objective observations) has been made yet.

Now comes the difficult part, namely this connection with the real world, i.e., with reproducible obejective observations in nature. In my opinion, the only consistent interpretation is the Minimal Statistical Interpretation, which is basically defined by Born's Rule, saying that for a given preparation of a system in a quantum state, represented by the Statistical Operator \hat{R} the probability (density) to measure a complete set of compatible operators is given by
P(A_1,\ldots, A_n|\hat{R})=\langle A_1,\ldots, A_n|\hat{R}|A_1,\ldots, A_n \rangle
where |A_1,\ldots, A_n\rangle is a (generalized) common eigenvector normalized to 1 (a \delta distribution) of the self-adjoint operators representing the complete set of compatible operators.

Now the interpretation is shifted to the interpretation of probabilities. QT makes no other predictions about the outcome of measurements than these probabilities, and now we have to think about the meaning of probabilities. It's clear that probability theory also is given as a axiomatic set of rules (e.g., the Kolmogorov axioms), which is unproblematic since its just a mathematical abstraction. The question now is, how to interpret probabilities in the sense of physical experiments. Physics is about the test of hypothesis about real-world experiments and thus we must make this connection between probabilities and outcomes of such real-world measurements. I don't see, how else you can define this connection than by repeating the measurement on a sufficiently large ensemble of identically and independently prepared experimental setups. The larger the ensemble the higher the statistical significance for proving or disproving the predicted probabilities for the outcome of measurements.

The Bayesian view, for me, is just a play with words, trying to give a physically meaningful interpretation of probability for a single event. In practice, however, you cannot prove anything about a probabilistic statement with only looking at a single event. If I predict a probability of 10% chance of rain tomorrow, and then the fact whether it rains or doesn't rain on the next day doesn't tell anything about the validity of my probabilistic prediction. The only thing one can say is that for many days with the weather conditions of today on average it will rain in 10% of all cases on the next day; no more no less. Whether it will rain or not on one specific date cannot be predicted by giving a probability.

So for the practice of physics the Bayesian view of probabilities is simply pointless, because doesn't tell anything about the outcome real experiments.
 
  • #13
vanhees71 said:
So for the practice of physics the Bayesian view of probabilities is simply pointless, because doesn't tell anything about the outcome real experiments.

I think it goes beyond physics. My background is in applied math and it invariably uses the frequentest interpretation which is basically the same as the ensemble interpretation. To me this Bayesian stuff seems just a play on words as well.

That said, and I can't comment because I didn't take those particular courses, applied Bayesian modelling and inference is widely taught - courses on it were certainly available where I went. I am not convinced however it requires the Bayesian interpretation.

Thanks
Bill
 
  • #14
vanhees71 said:
The Bayesian view, for me, is just a play with words, trying to give a physically meaningful interpretation of probability for a single event. In practice, however, you cannot prove anything about a probabilistic statement with only looking at a single event.

Well, the history of the universe only happens once, so we're stuck with having to reason about singular events, if we are to describe the universe.

More concretely, let's say that we have a theory that predicts that some outcome of an experiment has a 50/50 probability. So you perform the experiment 100 times, say, and find that outcome happens 49 times out 100. So that's pretty good. But logically speaking, how is making the conclusion based on 100 trials any more certain than making a conclusion based on 1 trial, or 10 trials? The outcome, 49 out of 100, is consistent with just about any probability at all. You haven't narrow down the range of probabilities at all. What have you accomplished, then? You've changed your confidence, or belief, that the probability is around 1/2.

Mathematically speaking, the frequentist account of probability is nonsense. Probability 1/2 doesn't mean that something will happen 1/2 of the time, no matter how many experiments you perform. And it's nonsensical to add "...in the limit as the number of trials goes to infinity...", also. There is no guarantee that relative frequencies approach any limit whatsoever.
 
  • #15
bhobba said:
I think it goes beyond physics. My background is in applied math and it invariably uses the frequentest interpretation which is basically the same as the ensemble interpretation. To me this Bayesian stuff seems just a play on words as well.

The frequentist interpretation really doesn't make any sense, to me. As a statement about ensembles, it doesn't make any sense, either. If you perform an experiment, such as flipping a coin, there is no guarantee that the relative frequency approaches anything at all in the limit as the number of coin tosses goes to infinity. Furthermore, since we don't really ever do things infinitely often, then what can you conclude, as a frequentist, from 10 trials of something. Or 100? Or 1000? You can certainly dutifully write down the frequency, but every time you do another trial, than number is going to change, by a tiny amount. Is the probability changing every time you perform the experiment?
 
  • #16
stevendaryl said:
The frequentist interpretation really doesn't make any sense, to me. As a statement about ensembles, it doesn't make any sense, either. If you perform an experiment, such as flipping a coin, there is no guarantee that the relative frequency approaches anything at all in the limit as the number of coin tosses goes to infinity. Furthermore, since we don't really ever do things infinitely often, then what can you conclude, as a frequentist, from 10 trials of something. Or 100? Or 1000? You can certainly dutifully write down the frequency, but every time you do another trial, than number is going to change, by a tiny amount. Is the probability changing every time you perform the experiment?

In practice, people who claim to be doing "frequentist probability" use "confidence intervals". So if you perform an experiment 100 times, and you get a particular outcome 49 times, then you can say something like: The probability is 49% +/- E, where E is a confidence interval. But it isn't really true. The "true" probability could be 99%. Or the "true" probability could be 1%. You could have just had a weird streak of luck. The choice of E is pretty much ad hoc.
 
  • #17
John Baez gives a discussion of Bayesianism here:
http://math.ucr.edu/home/baez/bayes.html

Here's a snippet:

It's not at all easy to define the concept of probability. If you ask most people, a coin has probability 1/2 to land heads up if when you flip it a large number of times, it lands heads up close to half the time. But this is fatally vague!

After all what counts as a "large number" of times? And what does "close to half" mean? If we don't define these concepts precisely, the above definition is useless for actually deciding when a coin has probability 1/2 to land heads up!

Say we start flipping a coin and it keeps landing heads up, as in the play Rosencrantz and Guildenstern are Dead by Tom Stoppard. How many times does it need to land heads up before we decide that this is not happening with probability 1/2? Five? Ten? A thousand? A million?

This question has no good answer. There's no definite point at which we become sure the probability is something other than 1/2. Instead, we gradually become convinced that the probability is higher. It seems ever more likely that something is amiss. But, at any point we could turn out to be wrong. We could have been the victims of an improbable fluke.

Note the words "likely" and "improbable". We're starting to use concepts from probability theory - and yet we are in the middle of trying to define probability! Very odd. Suspiciously circular.

Some people try to get around this as follows. They say the coin has probability 1/2 of landing heads up if over an infinite number of flips it lands heads up half the time. There's one big problem, though: this criterion is useless in practice, because we can never flip a coin an infinite number of times!

Ultimately, one has to face the fact that probability cannot be usefully defined in terms of the frequency of occurence of some event over a large (or infinite) number of trials. In the jargon of probability theory, the frequentist interpretation of probability is wrong.

Note: I'm not saying probability has nothing to do with frequency. Indeed, they're deeply related! All I'm saying is that we can't usefully define probability solely in terms of frequency.
 
  • #18
First, I don't know enough QM to have any opinion on interpretations of QM, but I do use Bayesian statistics in other things (e.g. analysis of medical tests)

vanhees71 said:
Well, maybe I'm just too much biased by my training as a physicist to make sense of the whole Baysian interpretation of probabilities. ...

So for the practice of physics the Bayesian view of probabilities is simply pointless, because doesn't tell anything about the outcome real experiments.
You should really invest a little time into it. The Bayesian approach to probaility is more in line with the scientific method than the frequentist approach.

In the scientific method you formulate a hypothesis, then you acquire data, then you use that data to decide to keep or reject your hypothesis. In other words, you want to determine the likelyhood of the hypothesis given the data, which is exactly what Bayesian statistics calculate. Unfortunately, frequentist statistical tests simply don't measure that. Instead they calculate the likelyhood of the data given the hypothesis.

I think that the big problem with Bayesian statistics right now is the lack of standardized tests. If you say "my t-test was significant with p=0.01" then everyone understands what mathematical test you ran on your data and what you got. There is no corresponding "Bayesian t-test" that you can simply report and expect everyone to know what you did.

Most likely, your preference for frequentist statistics is simply a matter of familiarity, born of the fact that the tools are well-developed and commonly-used. This seems to be the case for bhobba also.
bhobba said:
My background is in applied math and it invariably uses the frequentest interpretation which is basically the same as the ensemble interpretation.
 
  • #19
DaleSpam said:
You should really invest a little time into it. The Bayesian approach to probaility is more in line with the scientific method than the frequentist approach.

actually are merging. (frequentist and bayesian)

"Efron also compares more recent statistical theories such as frequentism to Bayes' theorem, and looks at the newly proposed fusion of Bayes' and frequentist ideas in Empirical Bayes. Frequentism has dominated for a century and does not use prior information, considering future behavior instead"

Read more at: http://phys.org/news/2013-06-bayesian-statistics-theorem-caution.html#jCp
 
  • #20
DaleSpam said:
First, I don't know enough QM to have any opinion on interpretations of QM, but I do use Bayesian statistics in other things (e.g. analysis of medical tests)

You should really invest a little time into it. The Bayesian approach to probaility is more in line with the scientific method than the frequentist approach.

I don't think that the frequentist interpretation of probability can be taken seriously, for the reasons that John Baez gives in the passage I quoted. On the other hand, a purely subjective notion of probability doesn't seem like the whole story, either.

For example, one could model a coin flip by using an unknown parameter h reflecting the probability of getting a "heads". One could start off with the completely unknown probability distribution on h: it could be anything between 0 and 1. Then you flip the coin a few times, and you use Bayes' theorem to get an adjusted probability distribution on the parameter h. For example, if I flip twice, and get 1 head and 1 tail, then the adjusted probability distribution is P(h) = 6h (1-h), which has a maximum at h=\frac{1}{2}.

The weird thing here is that you have probability appearing as an unknown parameter, h, and you also have it appearing as a subjective likelihood of that parameter. It doesn't make sense to me that it could all be subjective probability, because how can there be an unknown subjective probability h?
 
  • #21
stevendaryl said:
The frequentist interpretation really doesn't make any sense, to me. As a statement about ensembles, it doesn't make any sense, either. If you perform an experiment, such as flipping a coin, there is no guarantee that the relative frequency approaches anything at all in the limit as the number of coin tosses goes to infinity. Furthermore, since we don't really ever do things infinitely often, then what can you conclude, as a frequentist, from 10 trials of something. Or 100? Or 1000? You can certainly dutifully write down the frequency, but every time you do another trial, than number is going to change, by a tiny amount. Is the probability changing every time you perform the experiment?

The law of large numbers is rigorously provable from the axioms pf probability.

What it says is if a trial (experiment or whatever) is repeated a large number of times, independently under identical conditions, then the proportion of times that any specified outcome occurs approximately equals the probability of the event's occurrence on any particular trial; the larger the number of repetitions, the better the approximation tends to be.

This guarantees a sufficiently large, but finite, number of trials exists (ie an ensemble) that for all practical purposes contains the outcomes in proportion to its probability.

It seems pretty straightforward to me, but each to his own I suppose. This is applied math after all. I remember when I was doing my degree I used to get upset at careless stuff like treating dx as a small first order quantity - which of course it isn't - but things are often simpler doing that. Still the criticisms I raised then are perfectly valid and its only in a rigorous treatment they disappear - but things become a lot more difficult. If that's what appeals be my quest - I like to think I have come to terms with such things these days. As one of my statistics professors said to me - and I think it was the final straw that cured me of this sort of stuff - I can show you books where all the questions you ask are fully answered - but you wouldn't read them. He then gave me this deep tome on the theory of statistical inference - and guess what - he was right.

Thanks
Bill
 
  • #22
DaleSpam said:
Most likely, your preference for frequentist statistics is simply a matter of familiarity, born of the fact that the tools are well-developed and commonly-used. This seems to be the case for bhobba also.

That's true.

I too did a fair amount of statistics in my degree deriving and using stuff like the student t distribution, degrees of freedom etc. I always found the frequentest interpretation more than adequate.

That's not to say the Bayesian view is not valid - it is - its just I never found the need to move to that view.

Thanks
Bill
 
  • #23
bhobba said:
The law of large numbers is rigorously provable from the axioms pf probability.

What it says is if a trial (experiment or whatever) is repeated a large number of times, independently under identical conditions, then the proportion of times that any specified outcome occurs approximately equals the probability of the event's occurrence on any particular trial; the larger the number of repetitions, the better the approximation tends to be.

But the meaning of "tends to" is the part that makes no sense, under a frequentist account. What does that mean? It is possible, when flipping coins, to have a run of 1,000 flips in a row that are all heads. It is possible to have a run of 1,000,000 flips in a row with all heads. So what does this "tends to" mean? Well, you can say that such a run of heads is unlikely, but according to what meaning of "unlikely"?

This guarantees a sufficiently large, but finite, number of trials exists (ie an ensemble) that for all practical purposes contains the outcomes in proportion to its probability.

It doesn't guarantee anything. You can calculate a number, the likelihood that the relative frequency will differ from the probability by more than some specific amount. But what's the meaning of that number? It can't be given a frequentist interpretation as a probability.
 
  • #24
stevendaryl said:
But the meaning of "tends to" is the part that makes no sense, under a frequentist account. What does that mean? It is possible, when flipping coins, to have a run of 1,000 flips in a row that are all heads. It is possible to have a run of 1,000,000 flips in a row with all heads. So what does this "tends to" mean? Well, you can say that such a run of heads is unlikely, but according to what meaning of "unlikely"?

The meaning of such things lies in a rigorous development of probability. That's how it is proved and rests on ideas like almost surely convergence and convergence in probability.

You are putting the cart before the horse. In the frequentest interpretation you have an ensemble that is the EXPECTED outcome of a very large number of trials - and that's what the law of large numbers converges to. Sure - you can flip any number of heads - but that is not the expected value - which is for a large number half and half. Then one imagines a trial as picking a random element from that ensemble - you can pick any element of that ensemble every time - but the ensemble contains the objects in the correct proportion.

I agree these ideas are subtle and many great mathematicians such as Kolmorgorov wrote difficult tomes putting all this stuff on a firm basis eg the strong law of large numbers. But on a firm basis they most certainly are.

If you really want to investigate it without getting into some pretty hairy and advanced pure math then Fellers classic is a good place to start:
http://ruangbacafmipa.staff.ub.ac.id/files/2012/02/An-Introduction-to-probability-Theory-by-William-Feller.pdf

As you will see there is a lot of stuff that leads up to the proof of the law of large numbers and in volume 1 Feller does not give the proof of the very important Strong Law Of Large Numbers - he only gives the proof of the Weak Law - you have to go to volume 2 for that - and the level of math in that volume rises quite a bit.

Feller discusses the issues you raise and it is very subtle indeed - possibly even more subtle than you realize. However if that's what interests you then Feller is a good place to start.

Just as an aside I set myself the task of working through both volumes - bought them both. Got through volume 1 but volume 2 was a bit too tough and never did finish it.

Thanks
Bill
 
Last edited by a moderator:
  • #25
bhobba said:
I always found the frequentest interpretation more than adequate.
Yes, I have also, particularly with respect to the wide variety of powerful software with pre-packaged standard statistical tests. Personally, I think that the hard-core Bayesians need to spend less time promoting their viewpoint and more time developing and standardizing their tools.
 
  • #26
DaleSpam said:
Yes, I have also, particularly with respect to the wide variety of powerful software with pre-packaged standard statistical tests. Personally, I think that the hard-core Bayesians need to spend less time promoting their viewpoint and more time developing and standardizing their tools.

I am a bit more sanguine on it personally. I think either view is valid and the choice is a personal preference, its just, like you correctly said, the early probability courses at university, and certainly at high school, use the frequentest version almost exclusively so its a bit of a paradigm shift to move away from it.

I think most people are like me and want a real good reason to make the effort.

Thanks
Bill
 
  • #27
DaleSpam said:
First, I don't know enough QM to have any opinion on interpretations of QM, but I do use Bayesian statistics in other things (e.g. analysis of medical tests)

You should really invest a little time into it. The Bayesian approach to probaility is more in line with the scientific method than the frequentist approach.

In the scientific method you formulate a hypothesis, then you acquire data, then you use that data to decide to keep or reject your hypothesis. In other words, you want to determine the likelyhood of the hypothesis given the data, which is exactly what Bayesian statistics calculate. Unfortunately, frequentist statistical tests simply don't measure that. Instead they calculate the likelyhood of the data given the hypothesis.

I think that the big problem with Bayesian statistics right now is the lack of standardized tests. If you say "my t-test was significant with p=0.01" then everyone understands what mathematical test you ran on your data and what you got. There is no corresponding "Bayesian t-test" that you can simply report and expect everyone to know what you did.

Most likely, your preference for frequentist statistics is simply a matter of familiarity, born of the fact that the tools are well-developed and commonly-used. This seems to be the case for bhobba also.

I admit, I don't understand the Bayesian interpretation of probabilities. Particularly the subjectivity makes it highly suspicious for me.

In the natural sciences (and hopefully also in medicine and the social sciences) to the contrary, one has to try to make statements with the "least prejudice", given the (usually incomplete) information. One method is the maximum-entropy principle (based on information theory developed by Shannon and applied to (quantum) statistical physics by Jaynes). Such probabilities are subjective in the sense that you adapt your probability distribution to the (incomplete) information you have, but due to the maximum-entropy principle you use the distribution of least prejudice (in the sense of Shannon-Jaynes information theory).

In quantum theory, at least in the minimal statistical interpretation, probabilities are even "more objective" in the case of pure states. If you have prepared a system in a pure state, this means (from the theoretical point of view) that you have determined a complete set of compatible observables due to this preparation. Then quantum theory tells you precisely which probabilities for outcomes of further measurements you have to use (using the Heisenberg picture of time evolution for simplicity): The associated state of the system is given by the common eigenvector of the representing self-adjoint commuting operators of the complete set of compatible observables, which is uniquely defined up to a phase factor. The probability to measure any observable of the system at a later time is then given by the Born rule. There is no choice anymore which probability distribution you should associate with the situation. BTW. It's compatible with the maximum-entropy principle since given the values of the complete set of compatible observables it leads to the pure state, represented by the corresponding eigenvector.

For me the subjectivistic view of probabilities as seems to be used in Bayesian statistics is not scientific at all. Indeed, natural science is empirical, and you build theories which predict the outcome of measurements. These predictions have to be experimentally tested. Any experiment must be able to be reproducible precisely enough such that you can get "high enough statistics" to check a hypothesis quantitatively, i.e., to get the statistical significance of your measurement. Here, for simplicity, I assume an ideal experiment which has only negligible systematical errors.

All this uses the frequentist interpretation of probabilities, which is based on the law of large numbers, as bhobba already mentioned.

Perhaps it would help me to understand the Bayesian view, if you could explain how to test a probilistic theoretical statement empirically from this point of view. Is there a good book for physicists to understand the Bayesian point of view better?
 
  • #28
vanhees71 said:
For me the subjectivistic view of probabilities as seems to be used in Bayesian statistics is not scientific at all.

It scientific - just based on a different set of assumptions that's all.

What you have is three values of logic - true, false, and don't know - but don't know is assigned some kind of value indicating how close it is to true or false ie its plausibility. Its an extension of formal logic.

This value is assumed to obey some reasonable axioms - usually the so called Cox axioms. One of the key theorems is they imply the Kolmogorov axioms - with a minor difference - probabilities are finitely but not countably additive so one needs a slight further assumption to handle the infinite case.

As I indicated before the Kolmogrov axioms imply the frequentest interpretation via the law of large numbers which is provable in those axioms. What it says is, via some kind of probabilistic convergence, as the number of trials increases the number of each possible outcome approaches its expected number. Probabilistic convergence means the probability of it being that value gets higher and higher as it get closer and closer. This means, in the case of the law of large numbers, it is infinitesimally close 1 for a suitably large number. Now just like is done in many areas of applied math it is taken for all practical purposes to be 1 if it is infinitesimally close to it ie a dead cert it will have the proportion of outcomes the same as the probability.

What Stevendaryl seems to be concerned about is it is only infinitesimally close to it. As I say - each to their own I suppose - but I suspect virtually everyone that applies probability will accept it.

Thanks
Bill
 
  • #29
bhobba said:
The meaning of such things lies in a rigorous development of probability. That's how it is proved and rests on ideas like almost surely convergence and convergence in probability.

That's my point---frequentism can't be the basis of a rigorous development of probability. It can be an intuitive way to think about probability, but it doesn't actually work as a definition of probability.

You are putting the cart before the horse. In the frequentest interpretation you have an ensemble that is the EXPECTED outcome of a very large number of trials - and that's what the law of large numbers converges to.

Expected value is defined in terms of probability. So you can't use frequencies to be your definition of probability.
 
  • #30
bhobba said:
What Stevendaryl seems to be concerned about is it is only infinitesimally close to it. As I say - each to their own I suppose - but I suspect virtually everyone that applies probability will accept it.

"Infinitesimally close" is defined in terms of probability, but a notion of probability that is NOT relative frequency. So frequencies cannot be the core definition of probability.

I don't have any complaints with the use of frequencies as "frequentists" use them. It works for practical purposes, but it's not an actually consistent theory. In that sense, frequentism is sort of like the Copenhagen or "collapse" version of quantum mechanics. It's a set of rules for using probabilities, and it works well enough, but doesn't actually count as a rigorous theory.
 
  • #31
stevendaryl said:
"Infinitesimally close" is defined in terms of probability, but a notion of probability that is NOT relative frequency. So frequencies cannot be the core definition of probability.

I don't have any complaints with the use of frequencies as "frequentists" use them. It works for practical purposes, but it's not an actually consistent theory. In that sense, frequentism is sort of like the Copenhagen or "collapse" version of quantum mechanics. It's a set of rules for using probabilities, and it works well enough, but doesn't actually count as a rigorous theory.

Yes - by the Kolmogorov axioms. One of those axioms is the entire event space has probability 1. This means if an element of the event space is 1 it can be considered the entire event space. The law of large numbers shows for large n the outcomes are in proportion to the probabilities with probability so close to 1 it is for all practical purposes equal to it ie it is the entire event space.

If you don't consider that consistent then there are many areas in applied math where infinitesimal quantities are ignored so I guess you have trouble with those as well.

Thanks
Bill
 
  • #32
stevendaryl said:
That's my point---frequentism can't be the basis of a rigorous development of probability.

It isn't - the Kolmogorov axioms are. And that is what rigorous tretments use - not the frequentest interpretation. But the law of large numbers show for all practical purposes the frequentest interpretation is equivalent to it.

However I am taken back to what I was warned about all those years ago - you wouldn't read the tomes based on it - and in my experience that is very true. Even Fellers book, especially volume 2, is hard going, and believe me there are worse than that about.

Thanks
Bill
 
  • #33
Bayesianism vs Falsifiability

vanhees71 said:
Perhaps it would help me to understand the Bayesian view, if you could explain how to test a probilistic theoretical statement empirically from this point of view. Is there a good book for physicists to understand the Bayesian point of view better?

To me, Bayesianism provides a slightly different point of view about what it means to do science than the Karl Popper view that places "falsifiability" at the core.

In the Karl Popper view, we create a theory, we use the theory to make a prediction, we perform an experiment to test the prediction, and we discard the theory if the prediction is not confirmed. That's a nice, neat way of thinking about science, but it's an over-simplification, and it's incomplete. Why do I say that?

First, why is it an over-simplification? Because there is no way that a single experiment can ever falsify a theory. Whatever outcome happens from an experiment is capable of having multiple explanations. Some of those explanations imply that the theory used to make the prediction is simply wrong, and some do not. For example, if there is a glitch in a piece of equipment, you can't hold that against the theory. Also, every experiment that attempts to test a theory relies on interpretations that go beyond the theory. You don't directly measure "electron spin", for example, you measure the deflection of the electron in a magnetic field. So you need to know that there are no other forces at work on the electron that might deflect it, other than its magnetic moment (and you also need the theory connecting the electron's spin with its magnetic moment). So when a theory fails to make a correction, the problem could be in the theory, or it could be in the equipment, or it could be some random glitch, or it could be in the theory used to interpret the experimental result, or whatever. So logically, you can never absolutely falsify a theory. It gets even worse when the theories themselves are probabilistic. If the theory predicts something with 50% probability, and you observe that it has happened 49% of the time, has the theory been falsified, or not? There is no definite way to say.

Second, why do I say that falsifiability is incomplete? It doesn't provide any basis for decision-making. You're building a rocket, say, and you want to know how to design it so that it functions as you would like it to. Obviously, you want to use the best understanding of physics in the design of the rocket, but what does "best" mean? At any given time, there are infinitely many theories that have not yet been falsified. How do you pick out one as the "current best theory"? You might say that that's not science, that's engineering, but the separation is not that clear-cut, because, as I said, you need to use engineering in designing experimental equipment, and you have to use theory to interpret experimental results. You have to pick a "current best theory" or "current best engineering practice" in order to do experiments to test theories.

So how does Bayesianism come to the rescue? Well, nobody really uses Bayesianism in its full glory, because it's mathematically intractable, but it provides a model for how to do science that allows us to see what's actually done as a pragmatic short-cut.

In the Bayesian view, nothing besides pure mathematical claims is ever proved true or false. Instead, claims have likelihoods, and performing an experiment allows you to adjust those likelihoods.

So rather than saying that a theory is falsified by an experiment, the Bayesian would say that the theory's likelihood is decreased. And that's the way it really works. There was no single experiment that falsified Newtonian mechanics. Instead, there was a succession of experiments that cast more and more doubt on it. There was never a point where Newtonian mechanics was impossible to believe, it's just that at some point, the likelihood of Special Relativity and quantum mechanics rose to be higher (and by today, significantly higher) than Newtonian mechanics.

The other benefit, at least in principle, if not in practice, for Bayesianism is that it actually gives us a basis for making decisions about things like how to design rockets, even when we don't know for certain what theory applies. What you can do is figure out what you want to accomplish (get a man safely on the moon, for example), and try to maximize the likelihood of that outcome. If there are competing theories, then you include ALL of them in the calculation of likelihood. Mathematically, if O is the desired outcome, and E is the engineering approach to achieving it, and T_1, T_2, ... are competing theories, then

P(O | E) = \sum_i P(T_i) P(O | E, T_i)

You don't have to know for certain what theory is true to make a decision.
 
  • #34
bhobba said:
It isn't - the Kolmogorov axioms are. And that is what rigorous tretments use - not the frequentest interpretation. But the law of large numbers show for all practical purposes the frequentest interpretation is equivalent to it.

However I am taken back to what I was warned about all those years ago - you wouldn't read the tomes based on it - and in my experience that is very true. Even Fellers book, especially volume 2, is hard going, and believe me there are worse than that about.

Thanks
Bill

Okay, I guess if by "frequentism" you mean a particular methodology for using probabilities, then I don't have any big problems with it. But if it's supposed to explain the meaning of probabilities, I don't think it can actually do that, because you have to already have a notion of probability in order to connect relative frequencies to probabilities.
 
  • #35
stevendaryl said:
Okay, I guess if by "frequentism" you mean a particular methodology for using probabilities, then I don't have any big problems with it. But if it's supposed to explain the meaning of probabilities, I don't think it can actually do that, because you have to already have a notion of probability in order to connect relative frequencies to probabilities.

In modern times, just like many areas of mathematics, probability is defined in terms of axioms that are not given any meaning. Have a look at the Kolmogorov axioms - there is no meaning assigned to it whatsoever. The frequentest interpretation is one way to give it meaning - Baysian is another. This is the great strength of the formalist approach - a proof applies to however you relate it to stuff out there. Its great weakness is without the help of the intuition of an actual application you have to derive everything formally.

I agree justifying the frequentest interpretation without the Kolmogorov axioms leads to problems such as circular arguments. But then again I don't know of any book on probability that does that - all I have ever seen start with the Kolmogorov axioms and show, with varying degrees of rigor, and actually proving the key theorems, that the frequentest view follows from it.

Thanks
Bill
 
  • #36
vanhees71 said:
Perhaps it would help me to understand the Bayesian view, if you could explain how to test a probilistic theoretical statement empirically from this point of view.

Here's a simplified example. Suppose we have two competing theories about a coin: Theory A says that it is a fair coin, giving "heads" 1/2 of the time. Theory B says that it is a trick coin, weighted to give "heads" 2/3 of the time. To start off with, we don't have any reason for preferring one theory over the other, so we write:

P(A) = P(B) = \dfrac{1}{2}

Now flip the coin 4 times, and suppose you get HHTT. Call this event E. We compute probabilities:

P(E|A) = 0.0625

P(E|B) = 0.0494

P(E) = P(E|A) P(A) + P(E|B) P(B) = 0.0560

Now, the Bayesian rules say that we revise our likelihood of the two theories in light of this new information:

P'(A) = \dfrac{P(A) P(E|A)}{P(E)} = 0.558
P'(B) = \dfrac{P(B) P(E|B)}{P(E)} = 0.441

So based on this one experiment, the likelihood of theory A has risen, and the likelihood of B has fallen.
 
  • #37
bhobba said:
I agree justifying the frequentest interpretation without the Kolmogorov axioms leads to problems such as circular arguments. But then again I don't know of any book on probability that does that - all I have ever seen start with the Kolmogorov axioms and show, with varying degrees of rigor, and actually proving the key theorems, that the frequentest view follows from it.

I think maybe there's some disagreement about what "the frequentist view" is. If you mean that for many trials, the relative frequency gives you (with high probability) a good approximation to the probability, that's a conclusion from the axioms of probability, whether frequentist or bayesian. I thought that "the frequentist view" was that the meaning of probability is given by relative frequencies. That is not possible in a consistent way.
 
  • #38
stevendaryl said:
Here's a simplified example. Suppose we have two competing theories about a coin: Theory A says that it is a fair coin, giving "heads" 1/2 of the time. Theory B says that it is a trick coin, weighted to give "heads" 2/3 of the time. To start off with, we don't have any reason for preferring one theory over the other, so we write:

P(A) = P(B) = \dfrac{1}{2}

Now flip the coin 4 times, and suppose you get HHTT. Call this event E. We compute probabilities:

P(E|A) = 0.0625

P(E|B) = 0.0494

P(E) = P(E|A) P(A) + P(E|B) P(B) = 0.0560

Now, the Bayesian rules say that we revise our likelihood of the two theories in light of this new information:

P'(A) = \dfrac{P(A) P(E|A)}{P(E)} = 0.558
P'(B) = \dfrac{P(B) P(E|B)}{P(E)} = 0.441

So based on this one experiment, the likelihood of theory A has risen, and the likelihood of B has fallen.

I see, but that's nothing else than what I get with my "frequentist" approach. Here, you have a somewhat small ensemble of only 4 realizations of the experiment, but that's how I would do this statistical analysis as a "frequentist".
 
  • #39
vanhees71 said:
I see, but that's nothing else than what I get with my "frequentist" approach. Here, you have a somewhat small ensemble of only 4 realizations of the experiment, but that's how I would do this statistical analysis as a "frequentist".

I don't see that. What sense, in a frequentist approach, does it mean to say that theory A has probability 1/2 of being true, and theory B has a probability 1/2 of being true? That doesn't mean that half the time A will be true, and half the time B will be true.

I don't see that this example is compatible with frequentism, at all.
 
  • #40
vanhees71 said:
Particularly the subjectivity makes it highly suspicious for me.

In the natural sciences (and hopefully also in medicine and the social sciences) to the contrary, one has to try to make statements with the "least prejudice", given the (usually incomplete) information.
That can be done very easily in Bayesian statistics. Simply take what is called a non-informative prior, or an ignorance prior. Bayesian statistics and frequentist statistics are generally the same thing for a non-informative prior and a large amount of data.

However, Bayesian statistics let's you rationally account for information that you DO have in the form of an informed prior. Consider the recent FTL neutrino results from CERN, before the glitch was discovered. Most scientists looked at those results and rationally said something like "this new evidence is unlikely under SR, but we have all of this other evidence supporting SR so we still think that P(SR) is quite high even considering the new evidence, we await further information". That is a very Bayesian approach, and is the approach that rational people actually take when reasoning under uncertainty. When they have prior knowledge they integrate it into their evaluation of new evidence.

vanhees71 said:
Any experiment must be able to be reproducible precisely enough such that you can get "high enough statistics" to check a hypothesis quantitatively, i.e., to get the statistical significance of your measurement.
But this is exactly what you can not do with frequentist statistics. With frequentist methods you never test the hypothesis given the data, you always test the data given the hypothesis. When you do a frequentist statistical test the p value you obtain is the probability of the data, given the hypothesis. When doing science (at least outside of QM), most people think of the hypothesis as being the uncertain thing, not the data, but that is simply not what frequentist statistical tests measure.

vanhees71 said:
Perhaps it would help me to understand the Bayesian view, if you could explain how to test a probilistic theoretical statement empirically from this point of view. Is there a good book for physicists to understand the Bayesian point of view better?
I liked this series of video lectures by Trond Reitan:
http://www.youtube.com/playlist?list=PL066F123E80494F77

Be forwarned, it is a very low-budget production. He spends relatively little time on the philosophical aspects of Bayesian probability, but quite a bit of time on Bayesian inference and methods. I found it appealed to my "shut up and calculate" side quite a bit. The Bayesian methods and tests are scientifically more natural, regardless of how you choose to interpret the meaning of probability.
 
  • #41
stevendaryl said:
I think maybe there's some disagreement about what "the frequentist view" is. If you mean that for many trials, the relative frequency gives you (with high probability) a good approximation to the probability, that's a conclusion from the axioms of probability, whether frequentist or bayesian. I thought that "the frequentist view" was that the meaning of probability is given by relative frequencies. That is not possible in a consistent way.

I think our discussion has been slightly marred by a bit of a misunderstanding of what the other meant. I now see where you are coming from and agree. Basing probability purely on a frequentest interpretation has problems conceptually in that it can become circular. I suspect it can be overcome by a suitable amount of care - but why bother - the mathematically 'correct' way is via the Kolmogorov axioms and starting with that the frequentest interpretation is seen as a perfectly valid realization of those axioms based rigorously on the law of large numbers. Every book on probability I have read does it that way. Bayesian probability theory fits into exactly the same framework - although I personally haven't come across textbooks that do that but my understanding is they certainly exist, and in some areas of statistical inference may be a more natural framework. At least the university I went to certainly offers courses on it.

Thanks
Bill
 
Last edited:
  • #42
My take as a mathematicians is that a probability space is a type of mathematical structure just like a group or vector space or metric space. One wouldn't spend time arguing about whether this or that particular vector space is the real or more fundamental vector space, so why do it with probability spaces. Frequencies of results of repeatable experiments can be described by a probability space, so can a persons state of knowledge of factors contributing to the outcome of a single non-repeatable event. Probability is just a special case of the more general mathematical concept of a measure - a probability is a measure applied to parts of a whole indicating the relative extent the parts contribute to the whole in the manner under consideration. Saying something has a probability of 1/3 might mean that it came up 1/3 of the time in repeated experiment if that is what you are talking about (frequencies of outcomes) or it might mean that you know 30 scenarios 10 of which produce the outcome if that is what you a considering. Neither case is a more right or wrong example of probability, in the same way that neither SO(3) nor GL(4,C) is a more right or wrong use of group theory.
 
  • #44
vanhees71 said:
Perhaps it would help me to understand the Bayesian view, if you could explain how to test a probilistic theoretical statement empirically from this point of view. Is there a good book for physicists to understand the Bayesian point of view better?

I like Jaynes, Probability theory: The logic of science.
 
  • Like
Likes 1 person
  • #45
Have any of the QBist papers been accepted for publication? (As opposed to merely being uploaded to arxiv?) I'm not sure if I want to spend time reading lots of stuff that might turn out to be half-baked. (Various subtle puns in that sentence intended :) )
 
  • #46
Salman2 said:
Any comments (pro-con) on this Quantum Bayesian interpretation of QM by Fuchs & Schack ?: http://arxiv.org/pdf/1301.3274.pdf

I propose another variant of a "quantum Bayesian" interpretation, see arxiv.org:1103.3506

It is not completely Bayesian, instead, it is in part realistic, following de Broglie-Bohm about the reality of the configuration q(t). But it is Bayesian about the wave function.

Again, with care: What is interpreted as Bayesian is only the wave function of a closed system - that means, that of the whole universe. There is also the wave function we work with in everyday quantum mechanics. But this is only an effective wave function, It is defined, as in dBB theory, from the global wave function and the configuration of the environment, that means, mainly from the macroscopic measurement results of the devices used for the preparation of the particular quantum state.

Thus, because the configuration of the environment is ontic, the effective wave function is also defined by these ontic variables, thus, is essentially ontic. Therefore, no contradiction with the PBR theorem.

With this alternative in mind, I criticize QBism as following the wrong direction: Away from realism, away from nontrivial hypotheses about more fundamental theories. But this is what is IMHO the most important thing why it is important for scientists to think about interpretations at all. For computations, the minimal interpretation is sufficient. But it will never serve as a guide to find a more fundamenal theory.

This is different for dBB-like interpretations. They make additional hypotheses, about real trajectories q(t). Ok, we cannot test them now, and, because of the equivalence theorems, will be unable to test them in future too. A problem? Not really. Because we have internal problems of the interpretation itself, and these internal problems are a nice guide. We can try to find solutions for them, and these solutions may contain new, different physics, which becomes, then, testable.

This is also not empty talk. One internal problem of dBB are the infinities of the velocities \dot{q}(t) near the zeros of the wave function. Another one, related, is the Wallstom objection - the necessity to explain why probability and probabiliy flow combine into a wave function which appears if one does not consider the wave function as fundamental. To solve these problems, one has to make nontrivial assumptions about a subquantum theory, see arxiv.org:1101.5774. So, the interpretation gives strong hints where we have to look for physics different from quantum physics, in this case near the zeros of the wave function.

QBism, instead, does not lead to such hints where to look for new subquantum physics. The new mathematics of QBism looks like mathematics in the other, positivistic direction - not more but less restrictive, not less but more general. At least this is my impression.
 
  • #47
I just read for the first time the term QBism and found this discussion. On a first look, the paper by Fuchs and Schack looks horrible. Why pages full of speculations about what Feynman may have meant?
Isn't there some crisp axiomatic paper available?
 
  • #48
Last edited by a moderator:
  • #49
Mathematech said:
I've just come across this book http://www.springer.com/physics/the...+computational+physics/book/978-3-540-36581-5 based on Streater's website http://www.mth.kcl.ac.uk/~streater/lostcauses.html . I'm only just started reading it, seems that his views are totally against any notion of non-locality and that probability explains all the weirdness in QM. Comments?

The following says it all:
'This page contains some remarks about research topics in physics which seem to me not to be suitable for students. Sometimes I form this view because the topic is too difficult, and sometimes because it has passed its do-by date. Some of the topics, for one reason or another, have not made any convincing progress.'

There are many interpretations of QM - some rather 'backwater' like Nelson Stochastics. Some very mainstream and of great value in certain situations such as the Path Integral approach.

But as with any of them its pure speculation until someone can figure out an experiment to decide between them and have it carried out.

While discussion of interpretations in on topic in this forum, its kept on a tight leash to stop it degenerating into philosophy, which is off-topic.

So exactly what do you want to discuss - if you have in mind some interpretation, or issues in a specific interpretation, you want classification on, then fire away and me or others will see if they can help. Or do you want a general waffle about interpretations such as this doesn't tell us about realty (whatever that is - those that harp on it seldom define it - for good reason - its a philosophical minefield) or whatever which would be off topic.

Thanks
Bill
 
Last edited by a moderator:
  • #50
I want to discuss Streater's take that there is no need for assuming non-locality and that EPR etc can purely be understood via correct application of probability.
 

Similar threads

Replies
39
Views
4K
Replies
70
Views
1K
Replies
1
Views
2K
Replies
17
Views
1K
Replies
33
Views
4K
Replies
1
Views
2K
Back
Top