Derivation of the CHSH inequality

Alien8
Messages
77
Reaction score
1
Bell's 1971 derivation
The following is based on page 37 of Bell's Speakable and Unspeakable (Bell, 1971), the main change being to use the symbol ‘E’ instead of ‘P’ for the expected value of the quantum correlation. This avoids any implication that the quantum correlation is itself a probability.

We start with the standard assumption of independence of the two sides, enabling us to obtain the joint probabilities of pairs of outcomes by multiplying the separate probabilities, for any selected value of the "hidden variable" λ. λ is assumed to be drawn from a fixed distribution of possible states of the source, the probability of the source being in the state λ for any particular trial being given by the density function ρ(λ), the integral of which over the complete hidden variable space is 1. We thus assume we can write:

22652640b460de6399f0eb85fb5d75b5.png

where A and B are the average values of the outcomes. Since the possible values of A and B are −1, 0 and +1, it follows that:

d0823c0637b7ffae47eaa9232f8d0e97.png

Then, if a, a′, b and b′ are alternative settings for the detectors,

ba9e0adafbad6d0ef40507804a84790a.png

ff17109dfef0cfa366ccc63279ce1c5c.png

94f3fbf512dcca86f509d4b95b3c2aed.png

Then, applying the triangle inequality to both sides, using (5) and the fact that
03acf6967abe09c58288c5a22e32bf1e.png
as well as
8a8e5ce7571912da587c503c796d8446.png
are non-negative we obtain

ad55548f602bef99d7a667d8a5bdec53.png

64b0c82401c602b83937c84ff35c5674.png

or, using the fact that the integral of ρ(λ) is 1,

26a56c0d3c312cf050f75a4e3ade5126.png

which includes the CHSH inequality.
--- END QUOTE
http://en.wikipedia.org/wiki/CHSH_inequality#Derivation_of_the_CHSH_inequality1. "We start with the standard assumption of independence of the two sides, enabling us to obtain the joint probabilities of pairs of outcomes by multiplying the separate probabilities..."

Obtain the joint probability of what particular event?2. We see the premise from step (5) applied in step (6) which adds that \pm 1 into the equation, but without it, what is the underlying relation between four expectation values the equation describes? Would it be this: E(a,b) - E(a,b') = E(a,b) * E(a',b') - E(a,b') * E(a',b)?3. Step after (6), applying the triangle inequality to both sides. What is justification for this?
 
Last edited by a moderator:
Physics news on Phys.org
Step 6.
In that step, the idea they are using is that ##A = B + C## can also be written as ##A = B + C + D - D##. Basically we can add any term that is of the form ##D - D## since ##D - D = 0##.

There is E(a,b), E(a,b′), E(a′,b) and E(a′,b′), which I will call E1, E2, E3 and E4. Is that what you call A, B, C and D, so that: E1 = E2 + E3 + E4 - E4?
 
Alien8 said:
There is E(a,b), E(a,b′), E(a′,b) and E(a′,b′), which I will call E1, E2, E3 and E4. Is that what you call A, B, C and D, so that: E1 = E2 + E3 + E4 - E4?

The term I'm calling ##D## is ##\underline{A}(a,\lambda)\underline{B}(b,\lambda)\underline{A}(a',\lambda)\underline{B}(b',\lambda)##.

The term in the square brackets in the integrand is:
##\underline{A}(a,\lambda)\underline{B}(b,\lambda) - \underline{A}(a,\lambda)\underline{B}(b',\lambda)##
##= \underline{A}(a,\lambda)\underline{B}(b,\lambda) - \underline{A}(a,\lambda)\underline{B}(b',\lambda) + \underline{A}(a,\lambda)\underline{B}(b,\lambda)\underline{A}(a',\lambda)\underline{B}(b',\lambda) - \underline{A}(a,\lambda)\underline{B}(b,\lambda)\underline{A}(a',\lambda)\underline{B}(b',\lambda)##
##= \underline{A}(a,\lambda)\underline{B}(b,\lambda) - \underline{A}(a,\lambda)\underline{B}(b',\lambda) + \underline{A}(a,\lambda)\underline{B}(b,\lambda)\underline{A}(a',\lambda)\underline{B}(b',\lambda) - \underline{A}(a,\lambda)\underline{B}(b',\lambda)\underline{A}(a',\lambda)\underline{B}(b,\lambda)##
##= \underline{A}(a,\lambda)\underline{B}(b,\lambda) + \underline{A}(a,\lambda)\underline{B}(b,\lambda)\underline{A}(a',\lambda)\underline{B}(b',\lambda) - \underline{A}(a,\lambda)\underline{B}(b',\lambda) - \underline{A}(a,\lambda)\underline{B}(b',\lambda)\underline{A}(a',\lambda)\underline{B}(b,\lambda)##
##= \underline{A}(a,\lambda)\underline{B}(b,\lambda)[1 + \underline{A}(a',\lambda)\underline{B}(b',\lambda)] - \underline{A}(a,\lambda)\underline{B}(b',\lambda)[1+ \underline{A}(a',\lambda)\underline{B}(b,\lambda)]##
 
Last edited:
atyy said:
The term I'm calling ##D## is ##\underline{A}(a,\lambda)\underline{B}(b,\lambda)\underline{A}(a',\lambda)\underline{B}(b',\lambda)##.

The term in the square brackets in the integrand is:
##\underline{A}(a,\lambda)\underline{B}(b,\lambda) - \underline{A}(a,\lambda)\underline{B}(b',\lambda)##
##= \underline{A}(a,\lambda)\underline{B}(b,\lambda) - \underline{A}(a,\lambda)\underline{B}(b',\lambda) + \underline{A}(a,\lambda)\underline{B}(b,\lambda)\underline{A}(a',\lambda)\underline{B}(b',\lambda) - \underline{A}(a,\lambda)\underline{B}(b,\lambda)\underline{A}(a',\lambda)\underline{B}(b',\lambda)##
##= \underline{A}(a,\lambda)\underline{B}(b,\lambda) - \underline{A}(a,\lambda)\underline{B}(b',\lambda) + \underline{A}(a,\lambda)\underline{B}(b,\lambda)\underline{A}(a',\lambda)\underline{B}(b',\lambda) - \underline{A}(a,\lambda)\underline{B}(b',\lambda)\underline{A}(a',\lambda)\underline{B}(b,\lambda)##
##= \underline{A}(a,\lambda)\underline{B}(b,\lambda) + \underline{A}(a,\lambda)\underline{B}(b,\lambda)\underline{A}(a',\lambda)\underline{B}(b',\lambda) - \underline{A}(a,\lambda)\underline{B}(b',\lambda) - \underline{A}(a,\lambda)\underline{B}(b',\lambda)\underline{A}(a',\lambda)\underline{B}(b,\lambda)##
##= \underline{A}(a,\lambda)\underline{B}(b,\lambda)[1 + \underline{A}(a',\lambda)\underline{B}(b',\lambda)] - \underline{A}(a,\lambda)\underline{B}(b',\lambda)[1+ \underline{A}(a',\lambda)\underline{B}(b,\lambda)]##

So looking just at the equation in the first two lines, and given:

##\int \underline{A}(a,\lambda)\underline{B}(b,\lambda) \rho( \lambda ) d\lambda = E(a,b) = E1##
##\int \underline{A}(a,\lambda)\underline{B}(b',\lambda) \rho( \lambda ) d\lambda = E(a,b') = E2##
##\int \underline{A}(a',\lambda)\underline{B}(b',\lambda) \rho( \lambda ) d\lambda = E(a',b') = E3##
##\int \underline{A}(a',\lambda)\underline{B}(b,\lambda) \rho( \lambda ) d\lambda = E(a',b) = E4##

then in the original terms of expectation values it goes like this:

##E1 - E2 = E1-E2 + E1*E3 - E1*E4##

Correct? The question is where did that come from, according to what premise or mathematical principle is it supposed to be true? What is it E1, E2, E3 and E4 have in common to justify such a statement about their shared relationship? How do you prove that equation is true? Is it supposed to be true only for QM, only classical physics, or both?
 
Alien8 said:
So looking just at the equation in the first two lines, and given:

##\int \underline{A}(a,\lambda)\underline{B}(b,\lambda) \rho( \lambda ) d\lambda = E(a,b) = E1##
##\int \underline{A}(a,\lambda)\underline{B}(b',\lambda) \rho( \lambda ) d\lambda = E(a,b') = E2##
##\int \underline{A}(a',\lambda)\underline{B}(b',\lambda) \rho( \lambda ) d\lambda = E(a',b') = E3##
##\int \underline{A}(a',\lambda)\underline{B}(b,\lambda) \rho( \lambda ) d\lambda = E(a',b) = E4##

then in the original terms of expectation values it goes like this:

##E1 - E2 = E1-E2 + E1*E3 - E1*E4##

Correct?

No, that is not right. Don't think about the expectation values, just the term in the square brackets of the integrand.

##\underline{A}(a,\lambda)\underline{B}(b,\lambda) = D1##
##\underline{A}(a,\lambda)\underline{B}(b',\lambda) = D2##
##\underline{A}(a',\lambda)\underline{B}(b',\lambda) = D3##
##\underline{A}(a',\lambda)\underline{B}(b,\lambda) = D4##

Then the original term can be rewritten:

##D1 - D2 = D1-D2 + D1*D3 - D1*D3##
 
atyy said:
##\underline{A}(a,\lambda)\underline{B}(b,\lambda) = D1##
##\underline{A}(a,\lambda)\underline{B}(b',\lambda) = D2##
##\underline{A}(a',\lambda)\underline{B}(b',\lambda) = D3##
##\underline{A}(a',\lambda)\underline{B}(b,\lambda) = D4##

Then the original term can be rewritten:

##D1 - D2 = D1-D2 + D1*D3 - D1*D3##

Paraphrasing Alien8, one could ask the same question about D1, D2, D3, and D4:

The question is where did that come from, according to what premise or mathematical principle is it supposed to be true? What is it D1, D2, D3 and D4 have in common to justify such a statement about their shared relationship? How do you prove that equation is true? Is it supposed to be true only for QM, only classical physics, or both?
 
billschnieder said:
Paraphrasing Alien8, one could ask the same question about D1, D2, D3, and D4:

The question is where did that come from, according to what premise or mathematical principle is it supposed to be true?

We are back talking about relations between binary states instead of expectation values, even though they do not compare. Just like AB + AB' + A'B - A'B' = -2 or +2, that equation too is true according to pure algebra involving specifically numbers -1 and +1. There seem to be many combinations of arithmetic operations involving four variables with {-1,+1} limit that will yield equality, in which case the choice of that particular expression is simply arbitrary.

What is it D1, D2, D3 and D4 have in common to justify such a statement about their shared relationship?

The only thing they have in common is their limit {-1,+1}. But that is sufficient to construct numerous combinations of general algebraic equalities concerning four independent arbitrary variables with such a limit.

How do you prove that equation is true?

I evaluated the equation several times, each time assigning arbitrarily different -1 or +1 values to D variables. The equation produced 0 = 0, -2 = -2, and 2 = 2 results.

There seems to be an error in the third line though, which involves all four variables: D1-D2 = D1-D2 + D1*D3 - D2*D4 and it's true only for some combinations.

Is it supposed to be true only for QM, only classical physics, or both?

It's true for every possible combination of four variables under condition each one can be only -1 or +1. It's general and purely mathematical statement about numbers, it has no more to do with QM than 1 + 1 = 2.
 
Last edited:
Alien8 said:
There seems to be an error in the third line though, which involves all four variables: D1-D2 = D1-D2 + D1*D3 - D2*D4 and it's true only for some combinations.

Yes, if you assign the values D1,D2,D3,D4 with no relation between them (except that they have the same limits), then you will get an error in the third line. However, D1*D3 is not independent from D2*D4. In fact D1*D3 = D2*D4, because D1*D3 and D2*D4 are made up of the same "A" and "B" terms written in different orders.

So if you want to assign values between -1 and +1 independently to check the equation, you should assign them to the "A" and "B" terms.
 
atyy said:
No, that is not right. Don't think about the expectation values, just the term in the square brackets of the integrand.

##\underline{A}(a,\lambda)\underline{B}(b,\lambda) = D1##
##\underline{A}(a,\lambda)\underline{B}(b',\lambda) = D2##
##\underline{A}(a',\lambda)\underline{B}(b',\lambda) = D3##
##\underline{A}(a',\lambda)\underline{B}(b,\lambda) = D4##

If you lose the integral you don't have a number any more, but a binary state which is not a subject to arithmetic operations: dead & alive - alive & alive is undefined, it does not compute. That a photon goes one way or the other has no numerical value, it's an event or state. Assigning -1 and +1 labels to a binary state is very peculiar choice because it can obviously be misleading.

In any case, those D variables do not represent numbers, but four possible binary states or events: ##(++), (--), (+-), (-+)##. We can't do arithmetic with that, we need to count occurrences of many such events in order to work out probabilities and expectation values, and then we get the numbers we can actually do arithmetic with.

Then the original term can be rewritten:

##D1 - D2 = D1-D2 + D1*D3 - D1*D3##

Relation between independently arbitrary binary states is not relevant to expectation values. That two coins can either flip heads or tails has nothing to with how often they will both flip the same side. The derivation never looses the integral, expectation values E1 - E2 always remain on the left hand side of the equation. The equation ought to able to be expressed only in terms of expectation values, which is what the derivation begins and ends up with anyway.
 
  • #10
Alien8 said:
If you lose the integral you don't have a number any more, but a binary state which is not a subject to arithmetic operations
In hidden variable theory, you do have a number that is subject to arithmetic operations. That is the whole point of hidden variable theory.

If you deny that ##D1## is a definite number (which must be either ##+1## or ##-1##), then you are outside the framework of hidden variable theory, and the Bell inequalities cannot be derived.

The Bell inequalities apply only to local hidden variable theories, which are defined to be those theories in which ##A(a,\lambda)##, etc., have definite values (either ##+1## or ##-1##) for each value of the hidden variable ##\lambda##.

The Bell inequalities do not apply to any theory in which ##A(a,\lambda)## does not have a definite value (either ##+1## or ##-1##) for each value of the hidden variable ##\lambda##.

One such theory is quantum mechanics.
 
  • #11
Avodyne said:
If you deny that ##D1## is a definite number (which must be either ##+1## or ##-1##), then you are outside the framework of hidden variable theory, and the Bell inequalities cannot be derived.

I don't deny that, I observe that for probabilities and expectation values it is irrelevant whether the four possible events will be labeled ++, --, +-, -+ or 11, 00, 10, 01 or HH, TT, HT, TH, or whatever other binary state label with the Boolean domain. It's probability of those events happening which has a definite decimal range from 0.0 to 1.0, and it's expectation values which have definite decimal range from -1.0 to +1.0.

By the way, boolean logic operations do not directly translate to numbers arithmetic, and the outcome sample space always has the same boundary as input Boolean domain: {true, false}.

The Bell inequalities apply only to local hidden variable theories, which are defined to be those theories in which ##A(a,\lambda)##, etc., have definite values (either ##+1## or ##-1##) for each value of the hidden variable ##\lambda##.

The Bell inequalities do not apply to any theory in which ##A(a,\lambda)## does not have a definite value (either ##+1## or ##-1##) for each value of the hidden variable ##\lambda##.

The Bell inequality I quoted in the OP is not about binary states like: ##\underline{A}(a,\lambda)\underline{B}(b,\lambda)## , but about expectation values such as: ##\int \underline{A}(a,\lambda)\underline{B}(b,\lambda) \rho( \lambda ) d\lambda = E(a,b)##. Some people seems to think relations between binary state events are directly consequential to expectation values, but they are general and only define input domain limit, it's always the same for any theory you want to test the inequality against, and how the inequality will evaluate only depends on expectation value function or "hidden variable".
 
  • #12
Alien8 said:
The Bell inequality I quoted in the OP is not about binary states like: ##\underline{A}(a,\lambda)\underline{B}(b,\lambda)## , but about expectation values such as: ##\int \underline{A}(a,\lambda)\underline{B}(b,\lambda) \rho( \lambda ) d\lambda = E(a,b)##.
Sure, but you prove the Bell inequality for ##E(a,b)## by using properties of ##\underline{A}(a,\lambda)##.

Consider a particular combination of expectation values, ##E(a,b)+E(a,b')+E(a',b)-E(a',b')##. In local hidden variable theory, this combination can be written as

##E(a,b)+E(a,b')+E(a',b)-E(a',b')={}##
##\int\left[
\underline{A}(a,\lambda)\underline{B}(b,\lambda)
+\underline{A}(a,\lambda)\underline{B}(b',\lambda)
+\underline{A}(a',\lambda)\underline{B}(b,\lambda)
-\underline{A}(a',\lambda)\underline{B}(b',\lambda)\right]\!\rho(\lambda)d\lambda##

where ##\underline{A}(a,\lambda)##, etc., are each equal to ##+1## or ##-1##. (This is simply how we are choosing to represent the two binary values. This choice implies ##-1\le E(a,b)\le +1##.)

Do you agree with this, or not?
 
Last edited:
  • #13
Avodyne said:
Sure, but you prove the Bell inequality for ##E(a,b)## by using properties of ##\underline{A}(a,\lambda)##.

##\underline{A}(a,\lambda)## is unknown function with output sample space: {event 1, event 2}. This output is then only a part of the input for expectation value function. Event naming is arbitrary, probability only cares about the count or ratio of their occurrences.

where ##\underline{A}(a,\lambda)##, etc., are each equal to ##+1## or ##-1##. (This is simply how we are choosing to represent the two binary values. This choice implies ##-1\le E(a,b)\le +1##.)

Do you agree with this, or not?

Expectation value is measure of probabilities between four possible events. Probabilities naturally range from 0.0 to 1.0, so the reason why expectation values range from -1.0 to +1.0 is because: ##E(a,b) = P_{++}(a,b) + P_{--}(a,b) − P_{+-}(a,b) − P_{-+}(a,b)##.
 
  • #14
Alien8 said:
##\underline{A}(a,\lambda)## is unknown function with output sample space: {event 1, event 2}. This output is then only a part of the input for expectation value function. Event naming is arbitrary, probability only cares about the count or ratio of their occurrences.

You can use the concept of a random variable to assign numbers to the outcome.

http://en.wikipedia.org/wiki/Random_variable
http://www.stat.yale.edu/Courses/1997-98/101/ranvar.htm

Alien8 said:
Expectation value is measure of probabilities between four possible events. Probabilities naturally range from 0.0 to 1.0, so the reason why expectation values range from -1.0 to +1.0 is because: ##E(a,b) = P_{++}(a,b) + P_{--}(a,b) − P_{+-}(a,b) − P_{-+}(a,b)##.

The expectation value of a function of a random variable ##f(x)## is ##E(f(x)) = \int f(x) p(x) dx##. In general expectation values do not range between -1 and 1. However, if ##f(x)## ranges between -1 and 1, then the expectation ##E(f(x)) = \int f(x) p(x) dx## also ranges between -1 and 1.

http://mathworld.wolfram.com/ExpectationValue.html
 
  • #15
OK, then we can write

P_{++}(a,b)=\int_{\underline{A}(a,\lambda)=+1,\;\underline{B}(b,\lambda)=+1}\rho(\lambda)d\lambda

and similarly for ##P_{+-}(a,b)##, etc. This notation means that we do the integral only over those values of ##\lambda## for which both ##\underline{A}(a,\lambda)=+1## and ##\underline{B}(b,\lambda)=+1##. Here I am adopting the convention that the two binary results are called ##+1## and ##-1##. This convention gives us a notation that is useful, in the following sense: the outcomes ##++## and ##--## occur if and only if ##\underline{A}(a,\lambda)\underline{B}(b,\lambda)=+1##, and that the outcomes ##+-## and ##-+## occur if and only if ##\underline{A}(a,\lambda)\underline{B}(b,\lambda)=-1##. Therefore

P_{++}(a,b)+P_{--}(a,b) = \int_{\underline{A}(a,\lambda)\underline{B}(b,\lambda)=+1}\rho(\lambda)d\lambda
P_{+-}(a,b)+P_{-+}(a,b) = \int_{\underline{A}(a,\lambda)\underline{B}(b,\lambda)=-1}\rho(\lambda)d\lambda

Now, using ##E(a,b)=P_{++}(a,b)+P_{--}(a,b)-P_{+-}(a,b)-P_{-+}(a,b)##, we have

E(a,b)=\int_{\underline{A}(a,\lambda)\underline{B}(b,\lambda)=+1}\rho(\lambda)d\lambda<br /> -\int_{\underline{A}(a,\lambda)\underline{B}(b,\lambda)=-1}\rho(\lambda)d\lambda

Now comes the magic trick:

\int_{\underline{A}(a,\lambda)\underline{B}(b,\lambda)=+1}\rho(\lambda)d\lambda<br /> -\int_{\underline{A}(a,\lambda)\underline{B}(b,\lambda)=-1}\rho(\lambda)d\lambda<br /> =\int \underline{A}(a,\lambda)\underline{B}(b,\lambda)\rho(\lambda)d\lambda

You must stare at this until you understand it. It is the key to everything. The point is that the factor of ##\underline{A}(a,\lambda)\underline{B}(b,\lambda)## on the right-hand side takes on the values ##+1## or ##-1## only (because of the convention that we have adopted). When ##\underline{A}(a,\lambda)\underline{B}(b,\lambda)=+1##, we get the first term on the left-hand side, and when ##\underline{A}(a,\lambda)\underline{B}(b,\lambda)=-1##, we get the second term on the left-hand side.

I'll pause again. I'm hoping that I have now convinced you that, if we adopt the convention that the two binary values are ##+1## and ##-1##, then we can write

E(a,b)=\int \underline{A}(a,\lambda)\underline{B}(b,\lambda)\rho(\lambda)d\lambda
 
  • #16
atyy said:
However, if ##f(x)## ranges between -1 and 1, then the expectation ##E(f(x)) = \int f(x) p(x) dx## also ranges between -1 and 1.

Binary state events have no range, it's either one or the other. To calculate probabilities and expectation values it is irrelevant whether detections on the two detectors are marked with -1 and +1, or heads and tails. P(+1 and +1) can mean the same thing as P(heads and heads) if we choose so. The things inside probability function brackets are not numbers, but letters. Numbers arithmetic does not directly translate to logic operations of boolean algebra and probabilities.

Do you really mean to say if we decided to mark recordings of the two detectors with heads and tails instead of -1 and +1 the expectation value would range from heads to tails instead of from -1.0 to +1.0?
 
Last edited:
  • #17
Avodyne said:
...occur if and only if ##\underline{A}(a,\lambda)\underline{B}(b,\lambda)=-1##.

Why do you think you can multiply "photon A went left" with "photon B went right"?

I'll pause again. I'm hoping that I have now convinced you that, if we adopt the convention that the two binary values are ##+1## and ##-1##, then we can write E(a,b)=∫A(a,λ)B(b,λ)ρ(λ)dλ

The integral doesn't imply multiplication of the two terms, but pairing, enumeration and counting.
 
  • #18
Alien8 said:
Binary state events have no range, it's either one or the other.

Alien8, you are again starting to argue instead of trying to learn. What you should be taking away from the last few posts is that you will have to learn a bit more probability theory before you'll be ready to work through the CHSH derivation and proof.
 
  • #19
Alien8 said:
Binary state events have no range, it's either one or the other. ... Numbers arithmetic does not directly translate to logic operations of boolean algebra and probabilities.

You are right that discrete ("binary" is a special case of "discrete") outcomes don't have a range. However, we're talking about the expectation value of the result of series of such measurements, and that does have a range. Indeed, that's how casinos stay in business - every spin of the roulette wheel produces a discrete win-lose result for each bet, but the casino knows the exact expectation value of their winnings over a large number of such events, and it approaches a continuous function as the number of events becomes large.

In the Bell and CHSH experiments, the "correlation" values that appear in the formulas are all some form of ##(N_1-N_2)/(N_1+N_2)## where ##N_1## and ##N_2## are the number of trials in which both detectors gave the same result and the number of trials in which both detectors give different results. It should be clear that the expectation value of this quantity can take on values betwen -1 and 1, even though each individual trial has a binary result.
 
  • #20
Nugatory said:
Alien8, you are again starting to argue instead of trying to learn. What you should be taking away from the last few posts is that you will have to learn a bit more probability theory before you'll be ready to work through the CHSH derivation and proof.

I think the misunderstanding is about how the integral works and the meaning of the term ##\underline{A}(a,\lambda)\underline{B}(b,\lambda)## when separated out of this expression: ##
E(a,b)=\int \underline{A}(a,\lambda)\underline{B}(b,\lambda)\rho(\lambda)d\lambda##.Alain Aspect paper:
http://arxiv.org/abs/quant-ph/0402001

Equation (5):
##E(a,b) = P_{++}(a,b) + P_{--}(a,b) − P_{+-}(a,b) − P_{-+}(a,b)##

Equation (28):
##E(a,b) = \frac {N_{++}(a,b) + N_{--}(a,b) - N_{+-}(a,b) - N_{-+}(a,b)} {N_{++}(a,b) + N_{--}(a,b) + N_{+-}(a,b) + N_{-+}(a,b)}##I'd say these two equations make it pretty clear the two terms ##\underline{A}(a,\lambda)\underline{B}(b,\lambda)## are not multiplied under that integral, but paired and counted. It's not a pair of integers, it's not a pair of probabilities, it's a pair of events. Isn't that true?
 
Last edited:
  • #21
Alien8 said:
I'd say these two equations make it pretty clear the two terms ##\underline{A}(a,\lambda)\underline{B}(b,\lambda)## are not multiplied under that integral, but paired and counted. It's not a pair of integers, it's not a pair of probabilities, it's a pair of events. Isn't that true?

You aren't making much sense. The quantities appearing in the Bell correlator are the combinations of probabilities $$E_{xy} = P_{++}(xy) - P_{+-}(xy) - P_{-+}(xy) + P_{--}(xy) \,.$$ You can, of course, call the binary results anything you want. A common choice is to call them ##A_{x} = +1, -1## and ##B_{y} = +1, -1##. This is pure convention, but it's a very nice convention because it makes the combination of probabilities above is the same thing as the expectation value ##\langle A_{x} B_{y} \rangle## of the product of the random variables ##A_{x}## and ##B_{y}##. You can't multiply events, but you can certainly multiply the values of variables that you associate to those events, make predictions about their expectation values, and so on.
 
  • #22
wle said:
You aren't making much sense. The quantities appearing in the Bell correlator are the combinations of probabilities $$E_{xy} = P_{++}(xy) - P_{+-}(xy) - P_{-+}(xy) + P_{--}(xy) \,.$$

Please be specific what is not clear and how is what you said different than what I said.

## E(a,b) = \int \underline{A}(a,\lambda)\underline{B}(b,\lambda)\rho(\lambda)d\lambda = \frac {N_{++}(a,b) + N_{--}(a,b) - N_{+-}(a,b) - N_{-+}(a,b)} {N_{++}(a,b) + N_{--}(a,b) + N_{+-}(a,b) + N_{-+}(a,b)}##

Do you see right hand side is the result of A(a,λ) and B(b,λ) terms being multiplied under the integral, or being paired and counted to belong in one of the four possible combinations sets?
 
  • #23
wle said:
You can't multiply events, but you can certainly multiply the values of variables that you associate to those events, make predictions about their expectation values, and so on.

Please point some reference about the equation you are talking about. That two coins can either flip heads or tails has no bearing on how often they will both turn on the same side. It means that expectation value: E(a,b) = ∫A(a,λ)B(b,λ)ρ(λ)dλ is not relative or proportional to its input domain limit ##\{++, --, +-, -+\}##, which is always the same for every theory and every experiment, but it depends only on hidden variable λ.
 
  • #24
Alien8 said:
Please be specific what is not clear and how is what you said different than what I said.

## E(a,b) = \int \underline{A}(a,\lambda)\underline{B}(b,\lambda)\rho(\lambda)d\lambda = \frac {N_{++}(a,b) + N_{--}(a,b) - N_{+-}(a,b) - N_{-+}(a,b)} {N_{++}(a,b) + N_{--}(a,b) + N_{+-}(a,b) + N_{-+}(a,b)}##

Do you see right hand side is the result of A(a,λ) and B(b,λ) terms being multiplied under the integral, or being paired and counted to belong in one of the four possible combinations sets?

Huh? You estimate ##E_{xy}## in an experiment based on the detection counts ##N_{++}(x, y)## and so on. It is a hypothesis, introduced by Bell and motivated by reasoning about relativistic causality, that ##E_{xy}## should have a theoretical value of the form $$E_{xy} = \int \mathrm{d}\lambda \, \rho(\lambda) \, A_{x}(\lambda) B_{y}(\lambda) \,. \qquad (1)$$ That is what is being tested in a Bell experiment. You don't seem to have understood this. The point of a Bell experiment is to try to show that the assumption (1) is wrong.
 
  • #25
Alien8 said:
It means that expectation value: E(a,b) = ∫A(a,λ)B(b,λ)ρ(λ)dλ is not relative or proportional to its input domain limit ##\{++, --, +-, -+\}##, which is always the same for every theory and every experiment, but it depends only on hidden variable λ.

You are not making any sense at all here.
 
  • #26
wle said:
You are not making any sense at all here.

Please be specific what part is not clear and why. Input domain of A(a,λ) and B(b,λ) for the integral ∫A(a,λ)B(b,λ)ρ(λ)dλ is always the same {++,−−,+−,−+} for every theory and every experiment, yes? Thus the difference between expectation values predicted by different theories is proportional or depends only on hidden variable λ, no?

You estimate Exy in an experiment based on the detection counts N++(x,y) and so on. It is a hypothesis,

##E(a,b) = \frac {N_{++}(a,b) + N_{--}(a,b) - N_{+-}(a,b) - N_{-+}(a,b)} {N_{++}(a,b) + N_{--}(a,b) + N_{+-}(a,b) + N_{-+}(a,b)}##

It's not hypothesis, it's how expectation values is practically calculated, it's simply how expectation value looks naked without the integral. Counts ratio is just another way to express probabilities, it's the same equation as this: ##E(a,b) = P_{++}(a,b) + P_{--}(a,b) − P_{+-}(a,b) − P_{-+}(a,b)##.
 
  • #27
Alien8 said:
I think the misunderstanding is about how the integral works and the meaning of the term ##\underline{A}(a,\lambda)\underline{B}(b,\lambda)## when separated out of this expression: ##
E(a,b)=\int \underline{A}(a,\lambda)\underline{B}(b,\lambda)\rho(\lambda)d\lambda##.Alain Aspect paper:
http://arxiv.org/abs/quant-ph/0402001

Equation (5):
##E(a,b) = P_{++}(a,b) + P_{--}(a,b) − P_{+-}(a,b) − P_{-+}(a,b)##

Equation (28):
##E(a,b) = \frac {N_{++}(a,b) + N_{--}(a,b) - N_{+-}(a,b) - N_{-+}(a,b)} {N_{++}(a,b) + N_{--}(a,b) + N_{+-}(a,b) + N_{-+}(a,b)}##I'd say these two equations make it pretty clear the two terms ##\underline{A}(a,\lambda)\underline{B}(b,\lambda)## are not multiplied under that integral, but paired and counted. It's not a pair of integers, it's not a pair of probabilities, it's a pair of events. Isn't that true?

There are different ways of doing it, and also different people use the same words for different things, and different words for the same things. Wikipedia does CHSH in two different ways, and Alain Aspect does it yet another way. They are all correct (except maybe the reference to the triangle inequality by Wikipedia, which I don't understand). Anyway, if you want to stick to the derivation in the paper by Alain Aspect, that is a fine way to learn CHSH.
 
  • #28
atyy said:
There are different ways of doing it, and also different people use the same words for different things, and different words for the same things. Wikipedia does CHSH in two different ways, and Alain Aspect does it yet another way. They are all correct (except maybe the reference to the triangle inequality by Wikipedia, which I don't understand). Anyway, if you want to stick to the derivation in the paper by Alain Aspect, that is a fine way to learn CHSH.

Yes, I'd rather get back to my original question.

d0823c0637b7ffae47eaa9232f8d0e97.png


ba9e0adafbad6d0ef40507804a84790a.png

ff17109dfef0cfa366ccc63279ce1c5c.png

94f3fbf512dcca86f509d4b95b3c2aed.png


I'd like see how this equation in step (6) looks like without step (5) applied to it in terms of expectation values. So given:

∫A(a,λ)B(b,λ)ρ(λ)dλ = E(a,b)
∫A(a,λ)B(b′,λ)ρ(λ)dλ = E(a,b′)
∫A(a′,λ)B(b′,λ)ρ(λ)dλ = E(a′,b′)
∫A(a′,λ)B(b,λ)ρ(λ)dλ = E(a′,b)

...and without the insertion of +/- 1, does step (6) not boil down to this: E(a,b) - E(a,b') = E(a,b) * E(a',b') - E(a,b') * E(a',b)?
 
  • #29
Alien8 said:
Please be specific what part is not clear and why.

You are writing in a language I've never seen before. I don't know what you're calling an "input domain limit", what it means for an integral to "have" or be "relative" or "proportional" to an "input domain limit", or why you think it's true for every theory and every experiment.
##E(a,b) = \frac {N_{++}(a,b) + N_{--}(a,b) - N_{+-}(a,b) - N_{-+}(a,b)} {N_{++}(a,b) + N_{--}(a,b) + N_{+-}(a,b) + N_{-+}(a,b)}##

It's not hypothesis

##E(x, y) = \int \mathrm{d}\lambda \, \rho(\lambda) \, A_{x}(\lambda) B_{y}(\lambda)## is the hypothesis.
 
  • #30
wle said:
You are writing in a language I've never seen before. I don't know what you're calling an "input domain limit", what it means for an integral to "have" or be "relative" or "proportional" to an "input domain limit", or why you think it's true for every theory and every experiment.

- the domain of a function is the set of "input" or argument values for which the function is defined
http://en.wikipedia.org/wiki/Domain_of_a_function

- the sample space of an experiment or random trial is the set of all possible outcomes or results of that experiment
http://en.wikipedia.org/wiki/Sample_space

E(a,b) = ∫A(a,λ)B(b,λ)ρ(λ)dλ

1. outcome sample space for A(a,λ) and B(b,λ) integral pairs is always S = {++,−−,+−,−+}
2. sample space {++,−−,+−,−+} is thus the input domain for the integral ∫A(a,λ)B(b,λ)ρ(λ)dλ
3. 1 and 2 are true for every CHSH experiment and for every theory whether local or non-local
4. different theories predict different expectation values for E(a,b) given the same a and b
5. the difference between proposed E(a,b) functions is relative only to λ and not S = {++,−−,+−,−+}

##E(x, y) = \int \mathrm{d}\lambda \, \rho(\lambda) \, A_{x}(\lambda) B_{y}(\lambda)## is the hypothesis.

That's a description of expectation value in terms of hidden variable λ. The other one is the description in terms of probabilities. Neither are hypothetical, they are just algebraic. Hypothesis is E(x, y) = cos2(x-y), which just happens to be true, for some reason.
 
  • #31
Alien8 said:
- the domain of a function is the set of "input" or argument values for which the function is defined
http://en.wikipedia.org/wiki/Domain_of_a_function

2. sample space {++,−−,+−,−+} is thus the input domain for the integral ∫A(a,λ)B(b,λ)ρ(λ)dλ

No, that doesn't make sense, because the integral is not a function. Even if you recast it as a function, its inputs would be the whole functions ##A_{x}(\lambda)## and ##B_{y}(\lambda)##. It is not a function of ++, +-, -+, or --.
3. 1 and 2 are true for every CHSH experiment and for every theory whether local or non-local

That also doesn't make sense. ##\int \mathrm{d}\lambda \, \rho(\lambda) \, A_{x}(\lambda) B_{y}(\lambda)## is the expectation value for local theories only.
That's a description of expectation value in terms of hidden variable λ.

The hidden variable is hypothetical. We do not measure ##\lambda## in an experiment. We don't know that a ##\lambda## even exists at all.
Hypothesis is E(x, y) = cos2(x-y), which just happens to be true, for some reason.

No, that has been confirmed in experiments. That is not hypothetical.
 
  • #32
Alien8 said:
-
That [the integral defining ##E(x,y)##] is a description of expectation value in terms of hidden variable λ. The other one is the description in terms of probabilities. Neither are hypothetical, they are just algebraic

That integral is a calculation of the expectation value in terms of probabilities, under the hypothesis that the probabilities take a particular form. That hypothetical form makes them out to be functions of ##\lambda## and ##x## on one side, and ##\lambda## and ##y## on the other, but the appearance of ##\lambda## in the integral doesn't mean that this is somehow not an expression in terms of probabilities. Yes, the calculation is "just algebraic", but it's "just algebra" proceeding from the hypothesis, which is not:

Hypothesis is E(x, y) = cos2(x-y), which just happens to be true, for some reason.
No, that is an experimental observation, which just happens to be predicted by quantum mechanics. That observation also conflicts with the hypothesis above about the form of the probabilities.
 
  • #33
Alien8 said:
Yes, I'd rather get back to my original question.

d0823c0637b7ffae47eaa9232f8d0e97.png


ba9e0adafbad6d0ef40507804a84790a.png

ff17109dfef0cfa366ccc63279ce1c5c.png

94f3fbf512dcca86f509d4b95b3c2aed.png


I'd like see how this equation in step (6) looks like without step (5) applied to it in terms of expectation values. So given:

∫A(a,λ)B(b,λ)ρ(λ)dλ = E(a,b)
∫A(a,λ)B(b′,λ)ρ(λ)dλ = E(a,b′)
∫A(a′,λ)B(b′,λ)ρ(λ)dλ = E(a′,b′)
∫A(a′,λ)B(b,λ)ρ(λ)dλ = E(a′,b)

...and without the insertion of +/- 1, does step (6) not boil down to this: E(a,b) - E(a,b') = E(a,b) * E(a',b') - E(a,b') * E(a',b)?

No, including the +/- 1 it is

##E(a,b) - E(a,b') = \int[D1 - D2]\rho(\lambda)d\lambda = ∫[D1 - D2 + D1*D3 + D1*D3]\rho(\lambda)d\lambda##

where D1, D2, D3, D4 were defined in post #6.
 
  • #34
wle said:
No, that doesn't make sense, because the integral is not a function. Even if you recast it as a function, its inputs would be the whole functions ##A_{x}(\lambda)## and ##B_{y}(\lambda)##. It is not a function of ++, +-, -+, or --.

Would you please rather try to answer my question in post #29? Can you confirm what atty wrote above?
 
  • #35
atyy said:
No, including the +/- 1 it is

##E(a,b) - E(a,b') = \int[D1 - D2]\rho(\lambda)d\lambda = ∫[D1 - D2 + D1*D3 + D1*D3]\rho(\lambda)d\lambda##

where D1, D2, D3, D4 were defined in post #6.

Can we exclude step (5) and write the equation (6) without including +/- 1?Is ##E(a,b) - E(a,b') = \int[D1 - D2]\rho(\lambda)d\lambda## not the same thing as ##E(a,b) - E(a,b') = E(a,b) - E(a,b')##?In terms of D variables:

A(a,λ)B(b,λ) = D1
A(a,λ)B(b′,λ) = D2
A(a′,λ)B(b′,λ) = D3
A(a′,λ)B(b,λ) = D4

this right hand side of the step (6):
ff17109dfef0cfa366ccc63279ce1c5c.png

94f3fbf512dcca86f509d4b95b3c2aed.png


goes like this:
##\int D1[1 \pm D3 ] - \int D2[1 \pm D4 ]##?

Can we not just pull that ##1 \pm## straight out of it? How did that +/- 1 even got there, what kind of mathematics is that?
 
  • #36
Alien8 said:
Why do you think you can multiply "photon A went left" with "photon B went right"?
... The integral doesn't imply multiplication of the two terms, but pairing, enumeration and counting.
I tried to explain, very slowly and carefully, exactly how this notation works, and why multiplying ##A(a,\lambda)## by ##B(b,\lambda)## is equivalent to counting the number of matches minus the number of mismatches. But you don't seem to be willing to make any effort to understand why this is true, and instead just keep repeating the same stuff over and over.

So I'm out.
 
Last edited:
  • #37
Alien8 said:
Can we exclude step (5) and write the equation (6) without including +/- 1?Is ##E(a,b) - E(a,b') = \int[D1 - D2]\rho(\lambda)d\lambda## not the same thing as ##E(a,b) - E(a,b') = E(a,b) - E(a,b')##?In terms of D variables:

A(a,λ)B(b,λ) = D1
A(a,λ)B(b′,λ) = D2
A(a′,λ)B(b′,λ) = D3
A(a′,λ)B(b,λ) = D4

this right hand side of the step (6):
ff17109dfef0cfa366ccc63279ce1c5c.png

94f3fbf512dcca86f509d4b95b3c2aed.png


goes like this:
##\int D1[1 \pm D3 ] - \int D2[1 \pm D4 ]##?

Can we not just pull that ##1 \pm## straight out of it? How did that +/- 1 even got there, what kind of mathematics is that?

The ##1##s in ##[1 \pm D3 ] ## and ##[1 \pm D4 ] ## are what give you the ##2## in the final line of the CHSH derivation.
 
  • #38
wle said:
Huh? You estimate ##E_{xy}## in an experiment based on the detection counts ##N_{++}(x, y)## and so on. It is a hypothesis, introduced by Bell and motivated by reasoning about relativistic causality, that ##E_{xy}## should have a theoretical value of the form $$E_{xy} = \int \mathrm{d}\lambda \, \rho(\lambda) \, A_{x}(\lambda) B_{y}(\lambda) \,. \qquad (1)$$ That is what is being tested in a Bell experiment. You don't seem to have understood this. The point of a Bell experiment is to try to show that the assumption (1) is wrong.

For clarification : wle showed in post #16 how E(a,b) = int A (a,lambda)B(b,lambda)p(lambda)dlambda .
And Exy is the Null hypothesis with value < 2 that is to be dis-proven. Correct ?
 
Last edited:
  • #39
atyy said:
The ##1##s in ##[1 \pm D3 ] ## and ##[1 \pm D4 ] ## are what give you the ##2## in the final line of the CHSH derivation.

Step (6) doesn't only introduce +/- 1, there is a whole new equation in step (6) into which +/- 1 is inserted. I would like to know what that equation is and understand according to what principle +/- 1 gets to be where it ended up to be.

D1(1 ± D3) − D2(1 ± D4)

What is the expression before insertion of +/- 1 and why was it not inserted into the left hand side of the equation or next to D1 and D2 like this: (1 ± D1)(1 ± D3) − (1 ± D2)(1 ± D4)? What does it even mean, did you ever seen "±" in the middle of any equation before?
22652640b460de6399f0eb85fb5d75b5.png

where A and B are the average values of the outcomes. Since the possible values of A and B are −1, 0 and +1, it follows that:

At the very beginning it is already defined {-1, +1} is intrinsic limit to A and B. Expectation value E(a,b) already has its -1.0 to +1.0 range before any explicit insertion of heads and tails into the equation. It doesn't make any sense to take already existing intrinsic limits and then duplicate them explicitly in the same equation.
 
  • #40
Alien8 said:
Step (6) doesn't only introduce +/- 1, there is a whole new equation in step (6) into which +/- 1 is inserted. I would like to know what that equation is and understand according to what principle +/- 1 gets to be where it ended up to be.

D1(1 ± D3) − D2(1 ± D4)

What is the expression before insertion of +/- 1 and why was it not inserted into the left hand side of the equation or next to D1 and D2 like this: (1 ± D1)(1 ± D3) − (1 ± D2)(1 ± D4)? What does it even mean, did you ever seen "±" in the middle of any equation before?

The +/- is just short hand for two sets of equations.

The first set is:
##\begin{align}
E(a,b) - E(a,b') &= \int[D1 - D2]\rho(\lambda)d\lambda \\
&= ∫[D1 - D2 + D1*D3 - D1*D3]\rho(\lambda)d\lambda \\
&= ∫[D1 - D2 + D1*D3 - D2*D4]\rho(\lambda)d\lambda \\
&= ∫[D1 + D1*D3 - D2 - D2*D4]\rho(\lambda)d\lambda \\
&= ∫[D1*(1 + D3) - D2*(1 + D4)]\rho(\lambda)d\lambda \\
\end{align}##

The second set is:
##\begin{align}
E(a,b) - E(a,b') &= \int[D1 - D2]\rho(\lambda)d\lambda \\
&= ∫[D1 - D2 - D1*D3 + D1*D3]\rho(\lambda)d\lambda \\
&= ∫[D1 - D2 - D1*D3 + D2*D4]\rho(\lambda)d\lambda \\
&= ∫[D1 - D1*D3 - D2 + D2*D4]\rho(\lambda)d\lambda \\
&= ∫[D1*(1 - D3) - D2*(1 - D4)]\rho(\lambda)d\lambda \\
\end{align}##

 
  • Like
Likes Alien8
  • #41
atyy said:
The +/- is just short hand for two sets of equations.

The first set is:
##\begin{align}
E(a,b) - E(a,b') &= \int[D1 - D2]\rho(\lambda)d\lambda \\
&= ∫[D1 - D2 + D1*D3 - D1*D3]\rho(\lambda)d\lambda \\
&= ∫[D1 - D2 + D1*D3 - D2*D4]\rho(\lambda)d\lambda \\
&= ∫[D1 + D1*D3 - D2 - D2*D4]\rho(\lambda)d\lambda \\
&= ∫[D1*(1 + D3) - D2*(1 + D4)]\rho(\lambda)d\lambda \\
\end{align}##

Ok, I see now. Does the following hold true then as well:

##E(a,b) = \int D1 \rho(\lambda)d\lambda##
##E(a,b') = \int D2 \rho(\lambda)d\lambda##
##E(a',b') = \int D3 \rho(\lambda)d\lambda##
##E(a',b) = \int D4 \rho(\lambda)d\lambda##

##E(a,b) - E(a,b') = ∫[D1 + D1*D3 - D2 - D2*D4]\rho(\lambda)d\lambda##
##= ∫D1\rho(\lambda)d\lambda+ ∫D1\rho(\lambda)d\lambda*∫D3\rho(\lambda)d\lambda - ∫D2\rho(\lambda)d\lambda-∫D2\rho(\lambda)d\lambda*∫D4\rho(\lambda)d\lambda##
##= E(a,b) + E(a,b)*E(a',b') - E(a,b')-E(a,b')*E(a',b)##

##E(a,b) - E(a,b') = E(a,b) + E(a,b)*E(a',b') - E(a,b')-E(a,b')*E(a',b)##
 
Last edited:
  • #42
Alien8 said:
Ok, I see now. Does the following hold true then as well:

##E(a,b) = \int D1 \rho(\lambda)d\lambda##
##E(a,b') = \int D2 \rho(\lambda)d\lambda##
##E(a',b') = \int D3 \rho(\lambda)d\lambda##
##E(a',b) = \int D4 \rho(\lambda)d\lambda##

##E(a,b) - E(a,b') = ∫[D1 + D1*D3 - D2 - D2*D4]\rho(\lambda)d\lambda##
##= ∫D1\rho(\lambda)d\lambda+ ∫D1\rho(\lambda)d\lambda*∫D3\rho(\lambda)d\lambda - ∫D2\rho(\lambda)d\lambda-∫D2\rho(\lambda)d\lambda*∫D4\rho(\lambda)d\lambda##
##= E(a,b) + E(a,b)*E(a',b') - E(a,b')-E(a,b')*E(a',b)##

##E(a,b) - E(a,b') = E(a,b) + E(a,b)*E(a',b') - E(a,b')-E(a,b')*E(a',b)##

No because ##∫[D1*D3]\rho(\lambda)d\lambda \neq ∫D1\rho(\lambda)d\lambda*∫D3\rho(\lambda)d\lambda##.
 
  • #43
atyy said:
No because ##∫[D1*D3]\rho(\lambda)d\lambda \neq ∫D1\rho(\lambda)d\lambda*∫D3\rho(\lambda)d\lambda##.

Aren't they both constant for each a,a',b,b' combination so that: ##\int C * f(x)dx = C * \int f(x)dx## ?
 
  • #44
Can someone show the relationship and how E (a,b) and Exy are equated in the derivation ?
 
  • #45
Alien8 said:
Aren't they both constant for each a,a',b,b' combination so that: ##\int C * f(x)dx = C * \int f(x)dx## ?

No, because each ##D## term is not a constant, but does depend on ##\lambda## which is what you are integrating over. You can see this by looking at the definition of the ##D## terms in post #6.
 
  • #46
Avodyne said:
I have no idea what you mean. A local hidden variable theory says that the result of a spin measurement on a particle carrying hidden variable ##\lambda## when the detector is set to ##a## is given by a function ##A(a,\lambda)## that takes on the values ##+1## and ##-1## only (which is what we choose to call the two possible results).
Yes, but the inequalities involve expectation values not ##A(a, \lambda)##.

You said the counterfactual expectation values of the same set should be the same as the actual ones from a different set because hidden variable theories require that. And I'm saying it is wrong to think the combination of a mixture of actual and counterfactual expectation values from the same set is the same as the combination of actual expectation values from different sets, whether you have a hidden variable theory or not. The degree of freedom difference between two sets and 4 sets is not an issue that only applies to hidden variable theories. I gave an example of a local realistic coin for which the sum of actual and counterfactual expectation values was different from the sum of actual expectation values from two separate identical coins, and for each coin, the result I would have obtained by looking at the other side of the coin does exist.
 
  • #47
I've removed a number of posts that were taking us away from the original topic (how the CHSH inequality is derived). Please try to keep posts on topic here, and start a new thread if you want to challenge the derivation instead of helping alien8 understand it.
 
  • Like
Likes atyy
  • #48
Post #41 dispersed the thickest fog around this derivation for me, thanks atyy. I'm unhappy about binary states like heads and tails being labeled as integers and subjected to ordinary arithmetic, and I don't see expressions like black - white = 2 and white - black = -2 are meaningful, but at least I know now what's going on and can investigate further on my own.

Questions I have left concern only a single expectation value and these two curves:

300px-MalusQC.png

http://en.wikipedia.org/wiki/Local_hidden_variable_theory#Optical_Bell_tests

Q1.
I think the vertical axis is supposed to be marked "expectation value" instead of "correlation". Expectation value is: E = P(++) + P(--) - P(+-) - P(-+), that is ratio of matches - mismatches per total count, and therefore ranges from -1.0 to +1.0, for QM it's E = cos2(a-b). While correlation I think is only ratio of matches per total count or CORR = P(++) + P(--), therefore ranges from 0.0 to 1.0 and for QM it's CORR = cos^2(a-b). Right?

Q2.
http://en.wikipedia.org/wiki/CHSH_inequality#Derivation_of_the_CHSH_inequality : "We start with the standard assumption of independence of the two sides, enabling us to obtain the joint probabilities of pairs of outcomes by multiplying the separate probabilities..."

By "standard assumption of independence of the two sides" they mean "local theory" or "classical physics"? By "joint probabilities" they mean probabilities for each one of the four possible outcomes {++, --, +-, -+}, that is Pab(++), Pab(--), Pab(+-), Pab(-+)? And by "multiplying the separate probabilities" they mean P(A and B) = P(A)P(B), so that Pab(++) = Pa(+)Pb(+)?

Q3.
With common language established at Q1 and Q2, the actual question is how to obtain the integral or function which will plot each of those two curves. Let's start with the full line, a naive local prediction:

64701088dac3d49ea163ae2e72ddfbe0.png
,
http://en.wikipedia.org/wiki/Local_hidden_variable_theory#Optical_Bell_tests

If dotted curve is cos2(a-b), shouldn't full line be 1/2 cos2(a-b), how did they get 1/8 + 1/4 cos^(a-b)?

Q4.
Isn't it strange the local theory predicts proportionally varying correlation relative to two supposedly independent measurements? It's not as much as is observed apparently, but how in the world can a local theory conclude there would be any correlation between independent events at all?
 
  • #49
Nugatory said:
I've removed a number of posts that were taking us away from the original topic (how the CHSH inequality is derived). Please try to keep posts on topic here, and start a new thread if you want to challenge the derivation instead of helping alien8 understand it.

I apologize for a complaint that was based on my confusion. I was getting this thread confused with another thread on roughly the same topic.
 
  • #50
stevendaryl said:
No, a local theory doesn't imply independence of the results, and it does not imply P(A and B) = P(A)P(B). The reason why not is that even though A can't influence B, and B can't influence A, there might be a third cause that influences both. That's what the "local hidden variables" idea is all about: whether the correlations can be explained by assuming that there is a cause (the hidden variable) that influences both measurements.

A locally realistic model based on Malus' law is this: assume that in the twin-photon version of EPR, two photons are created with the same random polarization angle \phi. If Alice's filter is at angle \alpha then she detects a photon with probability cos^2(\alpha - \phi). Similarly, if Bob's filter is at angle \beta, then he detects a photon with probability cos^2(\alpha - \phi). The correlation E(\alpha, \beta) would then be:

E(\alpha, \beta) = P_{++} + P_{--} - P_{+-} - P_{-+}

where P_{++} is the probability both Alice and Bob detect a photon, P_{+-} is the probability Alice detects one and Bob doesn't, etc.

I read it like this: for the local theory based on Malus' law hidden variable ##\lambda## is just the common photon polarization ##\phi## shared by both photons in each entangled pair? Thus even though their interaction with separate analyzers are two independent events with separate independent probabilities P(A) and P(B), the ratio between their individual probabilities P(A) - P(B) would still be proportional to the ratio between their analyzer angle settings (a - b)?

For this model,
P_{++} = \frac{1}{\pi} \int d\phi cos^2(\alpha - \phi) cos^2(\beta - \phi) = \frac{1}{8} + \frac{1}{4}cos^2(\alpha - \beta)
P_{+-} = \frac{1}{\pi} \int d\phi cos^2(\alpha - \phi) sin^2(\beta - \phi) = \frac{1}{8} + \frac{1}{4}sin^2(\alpha - \beta)
P_{-+} = \frac{1}{\pi} \int d\phi sin^2(\alpha - \phi) cos^2(\beta - \phi) = \frac{1}{8} + \frac{1}{4}sin^2(\alpha - \beta)
P_{--} = \frac{1}{\pi} \int d\phi sin^2(\alpha - \phi) sin^2(\beta - \phi) = \frac{1}{8} + \frac{1}{4}cos^2(\alpha - \beta)

So
E(\alpha, \beta) = \frac{1}{2}(cos^2(\alpha - \beta) - sin^2(\alpha - \beta)) = \frac{1}{2} cos(2 (\alpha - \beta))

That's exactly 1/2 of the QM prediction.

So that equation I quoted from Wikipedia is just the probability for one of the four combinations, not really supposed to be marked as P(a,b), which I assumed is supposed to stand for E(a,b). That makes more sense.

I guess it doesn't make any difference, but why are you not integrating over 2pi or full 360 degrees? How do you convert that integral into the right hand side function for each of those probabilities, is there some simple principle behind it or you use some kind of integral calculator?
 

Similar threads

Replies
4
Views
2K
Replies
61
Views
4K
Replies
11
Views
2K
Replies
2
Views
1K
Replies
5
Views
3K
Replies
3
Views
707
Replies
5
Views
2K
Back
Top