Did they do a Loopholes free Bell test?

  • Thread starter Thread starter Nick666
  • Start date Start date
  • Tags Tags
    Bell Test
Click For Summary
The discussion centers on the status of loophole-free Bell tests, with participants noting that while some experiments have closed two out of three loopholes, a fully loophole-free test has not yet been achieved. Several research groups are reportedly planning or have begun such experiments, which are considered significant for quantum cryptography, particularly in establishing secure keys without trusting measurement devices. The conversation also touches on the complexities of closing all loopholes simultaneously, as conflicting requirements complicate experimental design. Participants express skepticism about the necessity of closing all loopholes, as many believe that existing violations of Bell's inequality will remain consistent regardless. Overall, the topic remains a critical area of inquiry in quantum physics, especially regarding its implications for theories of local realism and quantum mechanics.
  • #61
DirkMan said:
Since they used a quantum random number generator instead of the human random decisions, can no conclusions be drawn regarding this aspect ? Something like... we have as much "free will" as a quantum random number generator ? They do mention "free will" in the arvix paper .
IMHO, it would have been better to choose the settings by a cascade of classical (pseudo) random number generators. Or even just read them from a big file of pre-generated settings. But they need to get these random settings very fast and it appears that quantum RNGs are faster than state of the art classical pseudo RNGs.

Yes it is amusing that a Bell type experiment is supposed to show that nature is intrinsically non-deterministic (and moreover, in a non-local way) but to make the experiment convincing, we have to assume that we have at least effective local randomness. Thus you cannot escape the super-determinism (conspiracy) loophole. You have to make appeal to that loophole ludicrous. Occam's razor has to come to the rescue.
 
Physics news on Phys.org
  • #62
zonde said:
But I still don't get your argument against calling detection at C a post-selection.
With post-selection one usually means selection made after the experiment has been done, using knowledge of the results from both wings of the experiment.

Here, the selection is done outside the lightcone of the experiment. There is no way a local hidden variable model could make use of such a selection mechanism in order to violate the CHSH-inequality.
 
Last edited:
  • #63
zonde said:
Just to be sure about details I looked at the paper about their previous experiment - http://arxiv.org/abs/1212.6136

1. They do a lot of things to tune the setups at A and B and make photons indistinguishable.

2. Success of tuning is verified by observing HOM interference at C.

3. But I still don't get your argument against calling detection at C a post-selection.

4. And without singling out (post-selecting) just one of those entangled states there is no Bell inequality violation.
And the action that is occurring at C is interference between two photons (photon modes) so that |\psi^-\rangle and |\psi^+\rangle can be told apart.

Yes, there is post selection - no quibble about that. But if swapping were not occurring, that sample would not tell us anything. The local realist denies there is swapping that affects the resulting sample. They would say indistinguishability and overlap do not matter. Those things are ingredients for the swapping but NOT for the post-selection. That is what I am saying, as best I can tell.

1. For swapping purposes only. Not for post-selection.

2. HOM is not demonstrated on individual events. Just a couple of clicks separated by a specified timing.

3. As said, it is also post-selection.

4. I don't think so. The clicks in separate detectors show this, correct?

My assertion is that the post-selection steps could be performed WITHOUT indistinguishability (and therefore without swapping). Of course, then the experiment would not work. But the local realist shouldn't care since they will say that nothing that happens at C affects A & B's results. We know that idea is wrong.
 
  • #64
This beautifully crafted experiment now gives rise to a modified Quantum Randi Challenge (for all local realists out there):

Program three computers A, B, and C (or alternatively three subroutines that can be distributed on three computers), so that:

1. Computers A and B each send a signal to computer C.

2. Based on these signals, computer C outputs select/don't select.

3. Computers A and B are now both given exogenous binary random inputs, 0 or 1 (independently for both computers). 0 or 1 is just shorthand for each wing's binary choice of angles in a CHSH experiment.

4. Based on these inputs, computers A and B independently output 1 or -1.

5. Repeat from 1.

The only allowed communication between computers is the signals in step 1.

The challenge is this: For all selected pairs, the CHSH-inequality should be significantly violated. The above loop should run until the number of selected pairs is 1000 or more.
 
Last edited:
  • #65
Heinera said:
The above loop should run until the number of selected pairs is 1000 or more.
And why is 245 not enough?
 
  • #66
billschnieder said:
And why is 245 not enough?
Because we want to minimize the impact of flukes. You understand this, I'm sure.
 
  • Like
Likes billschnieder
  • #67
DrChinese said:
Yes, there is post selection - no quibble about that. But if swapping were not occurring, that sample would not tell us anything. The local realist denies there is swapping that affects the resulting sample. They would say indistinguishability and overlap do not matter. Those things are ingredients for the swapping but NOT for the post-selection.
Without post-selection, there will be no swapping. You have to understand that swapping is precisely the process of selecting a sub-ensemble which is correlated in a specific way from a larger ensemble which is not correlated.

For example, take a set of pairs of numbers X,Y where the corresponding numbers of each pair (x,y) are related by y = sin(x). It follows that x and y are correlated. With two such such sets, say X1, Y1, and X2, Y2 randomly generated at space-like separated locations A and B. We could do an additional "measurement" z on each x, e.g. z = cos(x), giving a set of measurement results Z1 and Z2 at A and B respectively. There is no correlation between X1 and X2 and therefore no correlation between Y1 and Y2 or Z1 and Z2. For each pair (x,y) from each arm we send the y value to a distant location C, even before the z values are available. At C we simply compare the incoming values and if they are indistinguishable, we generate a "good" signal. Based on the RNG used at A and B, we may not have very many "good" signals but we will get a few. By post-selecting the Z1 and Z2 values using the "good" results from location C, we get a sub-ensemble of the Z1 and Z2 results which are correlated with each other. This is essentially the process of "swapping". Replace "correlation" with "entanglement" and you have "entanglement swapping". Perhaps the confusion is with the common practice of discussing the technique in the context of a single measurement as opposed to ensembles.

There is no way for Alice(Bob) to know which of their results are "good" without information going from both Alice and Bob to station C, and then the back to Alice (Bob) in the form of "good" signals.

BTW, local realists do not deny swapping, they believe swapping has a local realistic explanation.
 
  • #68
billschnieder said:
Without post-selection, there will be no swapping. You have to understand that swapping is precisely the process of selecting a sub-ensemble which is correlated in a specific way from a larger ensemble which is not correlated.
Weather there is swapping or not does not really matter here. We could equally well have a theory where we just happened to produce the two electrons in an entangled state in the first place, and then the measurment of the photons at C would just confirm this entanglement, with no swapping taking place. The math would be the same. The only thing that matters for a LHV model, is that this confirmation is outside the lightcone of the experiment performed. So in no way can the settings or the results of the experiment influence the confirmation, nor vice versa.
 
  • #69
Heinera said:
Weather there is swapping or not does not really matter here.
It matters to have a proper understanding of what entanglement swapping entails.

The only thing that matters for a LHV model, is that this confirmation is outside the lightcone of the experiment performed. So in no way can the settings or the results of the experiment influence the confirmation, nor vice versa.
It matters also, that filtering results after the fact using information from both stations introduces "nonlocality". The "good" Z1 and Z2 ensembles are therefore nonlocally generated.
 
  • #70
billschnieder said:
It matters also, that filtering results after the fact using information from both stations introduces "nonlocality". The "good" Z1 and Z2 ensembles are therefore nonlocally generated.
But the whole point here is that in the experiment, they are not filtered after the fact . They are actually filtered prior to the fact (i.e. the performance of the experiment). See my post on the revised Quantum Randi Challenge earlier in this thread.
 
Last edited:
  • #71
Heinera said:
This beautifully crafted experiment now gives rise to a modified Quantum Randi Challenge (for all local realists out there):
I'm not a 'local realist' but I do simulations to test simple hypotheses and this has led me to conclude that the CHSH statistic cannot exceed the limit even with maximally correlated readings. Any simulation which does not include simething extra to mimic the entanglement will not break the limit.

From the 'socks' paper equ(13)

## E(a,b)=P(00|ab)+P(11|ab)-P(01|ab)-P(10|ab)##
##S=E(a,b)+E(a',b)+E(a,b')-E(a',b')##

so if ##P(00|ab)+P(11|ab)=0## (perfect anticorrelation) then ##S=-2##.

The (only ?) way to fake entanglement is to transform ##E(a,b)##
##E'(a,b)=2\sin(\theta)^2(P(00|ab)+P(11|ab))-2\cos(\theta)^2(P(01|ab)+P(10|ab)##

We must identify ##\theta## as a setting on the coincidence gathering apparatus. With this change the value of ##S## for a sample with zero mean correlation, S=2.0 for ##\theta=\pi/2##. If the correlation is not zero then S can break the limit. The picture shows S against ##\theta## on the x-axis from a sample that has a correlation of about 0.65 ( this is not a histogram it is ##\sin(x)^2##). The sample S value is 2.78 with SD=0.12 ( 100 runs of 1000 samples)

The justification for the cheat comes from the fact that entangling wave equations requires a change of Hilbert space basis ( eg a rotation) from the unentangled bases. The rotation used in the cheat comes from equation (4) in Bells paper.
 

Attachments

  • CHSH-a.png
    CHSH-a.png
    1.7 KB · Views: 408
Last edited:
  • #72
Heinera said:
With post-selection one usually means selection made after the experiment has been done, using knowledge of the results from both wings of the experiment.
That is precisely the case here. The measurements at Alice and Bob are post-selected after the experiment is done, using the "good" information from C.

Here, the selection is done outside the lightcone of the experiment. There is no way a local hidden variable model could make use of such a selection mechanism in order to violate the CHSH-inequality.
Not relevant to post-selection which is a non-local process irrespective of light-cones.
 
  • #73
DrChinese said:
Yes, there is post selection - no quibble about that. But if swapping were not occurring, that sample would not tell us anything. The local realist denies there is swapping that affects the resulting sample. They would say indistinguishability and overlap do not matter. Those things are ingredients for the swapping but NOT for the post-selection. That is what I am saying, as best I can tell.

1. For swapping purposes only. Not for post-selection.

2. HOM is not demonstrated on individual events. Just a couple of clicks separated by a specified timing.

3. As said, it is also post-selection.

4. I don't think so. The clicks in separate detectors show this, correct?

My assertion is that the post-selection steps could be performed WITHOUT indistinguishability (and therefore without swapping). Of course, then the experiment would not work. But the local realist shouldn't care since they will say that nothing that happens at C affects A & B's results. We know that idea is wrong.
First I would like to point out that the idea of entanglement swapping affecting measurement outcomes for entangled particles contradicts realism. Because if swapping happens in future light cones of two other measurements, swapping would have to affect measurements backwards in time.
So it's not about local realism but any realism.

Now let me give my argument why indistinguishability is required for successful post-selection. Here is quote from other paper about previous experiment http://arxiv.org/abs/1212.6136:
"The final state is one of two Bell states |\psi^\pm \rangle = 1/ \sqrt2 (|\uparrow_A\downarrow_B\rangle \pm |\downarrow_A\uparrow_B\rangle), with the sign depending on whether the same detector (+), or different detectors (−) clicked in the two rounds."

Without indistinguishability there would be a phase drift so that instead of two Bell states |\psi^\pm \rangle = 1/ \sqrt2 (|\uparrow_A\downarrow_B\rangle \pm |\downarrow_A\uparrow_B\rangle) we would get one classical state |\psi \rangle = 1/ \sqrt2 (|\uparrow_A\downarrow_B\rangle \otimes |\downarrow_A\uparrow_B\rangle) no matter if the same or different detectors click in the two rounds. And with that classical state we of course can't violate Bell inequalities.
 
  • #74
Heinera said:
Here, the selection is done outside the lightcone of the experiment. There is no way a local hidden variable model could make use of such a selection mechanism in order to violate the CHSH-inequality.
I am not claiming that local hidden variable model can violate the CHSH-inequality using that post-selection.
 
  • #75
billschnieder said:
That is precisely the case here. The measurements at Alice and Bob are post-selected after the experiment is done, using the "good" information from C.
*After* the experiment one has to gather together and correlate the information which has been generated at locations A, B and C. There is nothing wrong with selecting A and B data conditional on what was observed at C if the marker at C saying "go" was set before the randomisation of settings at A and B. See Bell (1981) "Bertlmann's socks", Figure 7 and the discussion of the experimental set-up around figure 7. The crucial point is do you accept that the settings are effectively random? Not available outside of their forward lightcones?
 
  • #76
gill1109 said:
*After* the experiment one has to gather together and correlate the information which has been generated at locations A, B and C.
Correct. Heinera seems to think otherwise. This experiment is not an "event-ready" experiment, in which Alice and Bob are told "go" each iteration. They measure everything and the results are post-selected *after*.

Post selection introduces nonlocality and nonfactorability in the data.
 
  • #77
billschnieder said:
Correct. Heinera seems to think otherwise. This experiment is not an "event-ready" experiment, in which Alice and Bob are told "go" each iteration. They measure everything and the results are post-selected *after*.

Post selection introduces nonlocality and nonfactorability in the data.
No, I do not think otherwise. The decision to select or not is made before the experiment is performed. If the decision is to not select, what difference does it make if Alice and Bob then perform the experiment or not, when it is already decided that the result will not be used?
 
  • #78
Heinera said:
No, I do not think otherwise. The decision to select or not is made before the experiment is performed. If the decision is to not select, what difference does it make if Alice and Bob then perform the experiment or not, when it is already decided that the result will not be used?
Of course the experiment was designed with the intention to select. But the selection is done *after* the experiment not before. Post-selection makes the difference that the resulting *post-selected* ensemble is nonlocal. Alice and Bob do not know at each instant if they should "go" or not. They always do the measurement and only at the *end*, do they reject the "bad" results after the information from C has been communicated to them. That's all I'm saying.
 
  • #79
billschnieder said:
Of course the experiment was designed with the intention to select. But the selection is done *after* the experiment not before.

Post-selection makes the difference that the resulting *post-selected* ensemble is nonlocal. That's all I'm saying.
Then you don't understand the experiment. The decision to select or not is done (and recorded) before each experiment is performed.
 
  • Like
Likes Nugatory
  • #80
Heinera said:
Then you don't understand the experiment. The decision to select or not is done (and recorded) before each experiment is performed.
Exactly. The time at which you analyse the data is irrelevant. All CHSH experiments involve bringing data together *after* the experiment and selecting four subsets according to the settings at the two locations. This selection necessarily is done in a non-local way and after the experiment. But so what? The way the data is selected does not bias the estimated correlation. That's the important thing.
 
  • #81
Heinera said:
Then you don't understand the experiment. The decision to select or not is done (and recorded) before each experiment is performed.
Perhaps because you are using a very strange meaning of "experiment". The experiment is not just what happens in one iteration. It is certainly a fact that post-selection is done *after* the experiment, which means all the runs have been completed and all the measurements at Alice and Bob have been done. It is also a fact that such post-selection, introduces nonlocality such that the ensemble of results left after filtration is not factorable. This is obvious. This is what I'm saying and I don't think you disagree with that.

The article reports the result of one experiment that involved many iterations. Not 245 separate experiments. Actually, by your definition of "experiment", they did many millions of experiments, and presented the results of just 245 of them. If the "go" signal was available before the "experiment", why waste time to do all the "bad" ones? The reported violation is from jointly considering all 245 "good" results. The fact that there is no non-locality in a single run (which you call "experiment") does not change the fact that as a whole, the final ensemble of "good" results was post-selected and therefore nonlocal.
 
Last edited:
  • #82
billschnieder said:
Perhaps because you are using a very strange meaning of "experiment". The experiment is not just what happens in one iteration.
Yes, with "experiment" I mean what happens in one iteration. Each iteration is an individual experiment.
billschnieder said:
It is certainly a fact that post-selection is done *after* the experiment, which means all the runs have been completed and all the measurements at Alice and Bob have been done.
This is only for convenience. They could just as well analyze the data in real time, and get an increasingly more significant violation as the experiments progressed. The end result would be the same.
billschnieder said:
It is also a fact that such post-selection, introduces nonlocality such that the ensemble of results left after filtration is not factorable. This is obvious. This is what I'm saying and I don't think you disagree with that.
No, it's not obvious. In fact, it's wrong, so I do disagree with that.
billschnieder said:
The article reports the result of one experiment that involved many iterations. Not 245 separate experiments. Actually, by your definition of "experiment", they did many millions of experiments, and presented the results of just 245 of them. If the "go" signal was available before the "experiment", why waste time to do all the "bad" ones?
They were not wasting any time. In order to create the 245 entangled states, they had to make millions of tries. It was those tries that took the time, and they were unavoidable. Whether an experiment was or was not performed after those tries didn't imply any wasting of time (because in order to not perform the experiment, A and B would have to waste time waiting for a return signal from C that said "no-go").
 
  • #83
Heinera said:
Yes, with "experiment" I mean what happens in one iteration. Each iteration is an individual experiment.
That's not what I meant by experiment, so you misunderstood what I said.

Heinera said:
This is only for convenience. They could just as well analyze the data in real time, and get an increasingly more significant violation as the experiments progressed. The end result would be the same.
Nope they can't, there is no way for Alice (Bob) to know what "experiment" (your definition) is good in real time. Even if there was a way, there is no way to analyze any data in real time. Analysis requires the results from Alice and Bob to be brought together *after* the all the runs have been completed. This is obvious.

Heinera said:
No, it's not obvious. In fact, it's wrong, so I do disagree with that.
That is because you don't understand it. It simply is the fact that you need information from both Alice and Bob to filter the results at Alice, as well as the results at Bob. Therefore the filtered results are nonlocal. What is wrong in that? There is no way for Alice to know which of her results are "good" until the information she sends through the photon is compared with the information sent by Bob, and vice versa. it is as clear as a bell that this is nonlocal filtration.

Heinera said:
They were not wasting any time. In order to create the 245 entangled states, they had to make millions of tries.
I did not say they were wasting time. I said if as you suggested earlier, the "event-ready" information was available before the "experiment" (your definition), then it makes no sense for Alice(Bob) to measure the "bad" results -- that would be a waste of time. That they waited until *after* the measurement to filter out the "good" from the "bad" owes to the fact that the "good" signal was only available *after* the measurements had been made, which was my point all along. That is why it is *post*-selection.
 
Last edited:
  • #84
billschnieder said:
Nope they can't, there is no way for Alice (Bob) to know what "experiment" (your definition) is good in real time. Even if there was a way, there is no way to analyze any data in real time. Analysis requires the results from Alice and Bob to be brought together *after* the all the runs have been completed. This is obvious.
Alice and Bob don't need to know what experiment is good in real time, because it is not Alice and Bob that is doing the analysis. They are not real human beings, you know. Of course analysis can be done in real time before all the runs have been completed. It's just that as more runs come in, the analysis becomes more and more significant.

billschnieder said:
That is because you don't understand it. It simply is the fact that you need information from both Alice and Bob to filter the results at Alice, as well as the results at Bob. Therefore the filtered results are nonlocal. What is wrong in that? There is no way for Alice to know which of her results are "good" until the information she sends through the photon is compared with the information sent by Bob, and vice versa. it is as clear as a bell that this is nonlocal filtration.
Now it's clear that you don't understand the experiment. No information from the results at Alice or Bob is needed to filter the runs in this experiment. Only information from C is needed, which is available even before Alice and Bob have performed their run.
 
Last edited:
  • #85
billschnieder said:
I did not say they were wasting time. I said if as you suggested earlier, the "event-ready" information was available before the "experiment" (your definition), then it makes no sense for Alice(Bob) to measure the "bad" results -- that would be a waste of time. That they waited until *after* the measurement to filter out the "good" from the "bad" owes to the fact that the "good" signal was only available *after* the measurements had been made, which was my point all along. That is why it is *post*-selection.

Heinera said:
Now it's clear that you don't understand the experiment. No information from the results at Alice or Bob is needed to filter the runs in this experiment. Only information from C is needed, which is available even before Alice and Bob have performed their run.

According to the text and Figure 2 of the ArXiv preprint [arXiv:1508.05949], the "go/no go" determination event is spacelike separated from Alice's and Bob's measurement choices and recording of outcomes, so the ##C = \text{go/no go}## outcome should neither causally influence nor be influenced by Alice's and Bob's measurement choices or results.

As far as Bell's theorem is concerned, at least one way you can formally accommodate the postselection is by considering it part of a three-party Bell scenario. Under the locality hypothesis, this means that the (single-round) joint probability distribution should factorise according to $$P(abc \mid xy) = \int \mathrm{d}\lambda \, \rho(\lambda) \, P_{\mathrm{A}}(a \mid x; \lambda) \, P_{\mathrm{B}}(b \mid y; \lambda) \, P_{\mathrm{C}}(c \mid \lambda) \,,$$ where ##a, b, x, y## are Alice's and Bob's inputs and outputs and ##c \in \{\text{go}, \text{no go}\}## is the event-ready outcome. From there, high school-level probability theory will tell you that conditioning on ##c = \text{go}## does not allow you to fake a nonlocal probability distribution between Alice and Bob's systems: $$\begin{eqnarray*}
P(ab \mid xy; c = \text{go}) &=& \frac{P(ab, c = \text{go} \mid xy)}{P(c = \text{go} \mid xy)} \\
&=& \int \mathrm{d}\lambda \, \rho(\lambda) \, \frac{P_{\mathrm{C}}(c = \text{go} \mid \lambda)}{P(c = \text{go})} \, P_{\mathrm{A}}(a \mid x; \lambda) \, P_{\mathrm{B}}(b \mid y; \lambda) \\
&=& \int \mathrm{d}\lambda \, \rho(\lambda \mid c = \text{go}) \, P_{\mathrm{A}}(a \mid x; \lambda) \, P_{\mathrm{B}}(b \mid y; \lambda) \,,
\end{eqnarray*}$$ which uses the no-signalling conditions (implied by locality) to say that ##P(c = \text{go} \mid xy) = P(c = \mathrm{go})## is independent of ##x## and ##y## and Bayes' theorem to say that ##\rho(\lambda \mid c = \text{go}) = \rho(\lambda) \, \frac{P_{\mathrm{C}}(c = \text{go} \mid \lambda)}{P(c = \text{go})}##.

The end result, with the conditioning on "go", has the same local factorisation that is used in the (single round) derivation of Bell inequalities. (And from there, statistical analyses of the sort done by Gill and others explain how you can turn this into a hypothesis test applicable to a real multi-round Bell experiment.)
 
Last edited:
  • Like
Likes Mentz114
  • #86
billschnieder said:
Without post-selection, there will be no swapping. You have to understand that swapping is precisely the process of selecting a sub-ensemble which is correlated in a specific way from a larger ensemble which is not correlated.

For example, take a set of pairs of numbers X,Y where the corresponding numbers of each pair (x,y) are related by y = sin(x). It follows that x and y are correlated. With two such such sets, say X1, Y1, and X2, Y2 randomly generated at space-like separated locations A and B. We could do an additional "measurement" z on each x, e.g. z = cos(x), giving a set of measurement results Z1 and Z2 at A and B respectively. There is no correlation between X1 and X2 and therefore no correlation between Y1 and Y2 or Z1 and Z2. For each pair (x,y) from each arm we send the y value to a distant location C, even before the z values are available. At C we simply compare the incoming values and if they are indistinguishable, we generate a "good" signal. Based on the RNG used at A and B, we may not have very many "good" signals but we will get a few. By post-selecting the Z1 and Z2 values using the "good" results from location C, we get a sub-ensemble of the Z1 and Z2 results which are correlated with each other. This is essentially the process of "swapping". Replace "correlation" with "entanglement" and you have "entanglement swapping". Perhaps the confusion is with the common practice of discussing the technique in the context of a single measurement as opposed to ensembles.

There is no way for Alice(Bob) to know which of their results are "good" without information going from both Alice and Bob to station C, and then the back to Alice (Bob) in the form of "good" signals.

BTW, local realists do not deny swapping, they believe swapping has a local realistic explanation.

Sorry, you DO deny that swapping is an essential requirement to CAUSING the entanglement. You equate swapping to detection (by definition), and I am saying you can detect without swapping.

Entanglement swapping = post-selection (B&C) + indistinguishability (B&C).

How do they check indistinguishability (B&C)? By looking for entanglement on A & D using the post selection group ONLY! There cannot be a local causal connection between the swapping and the A & D entanglement!

In other words: The local realist will have to admit that there is nothing occurring at A & D based on a decision made at B & C. But I can choose to post-select the A & D pairs using the B & C criteria WITH or WITHOUT having the B & C pairs be indistinguishable.

For example, I could place a linear polarizer in front of B and/or C. That should mean nothing to the local realist, since they believe the linear polarization is already determined. But for QM, that means that the photons B and C are distinguishable even though they otherwise will meet the post-selection criteria. There will be no entanglement at A & D, no violation of a Bell inequality.

Entanglement swapping is a physical process and is NOT just post-selection. It does require post-selection to determine whether swapping successfully occurred. The physical process (entanglement swapping) causes entanglement at A & D, demonstrating quantum non-locality.
 

Similar threads

Replies
4
Views
1K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 82 ·
3
Replies
82
Views
11K
Replies
19
Views
3K
  • · Replies 8 ·
Replies
8
Views
2K
Replies
6
Views
3K
  • · Replies 75 ·
3
Replies
75
Views
12K
Replies
63
Views
8K
  • · Replies 71 ·
3
Replies
71
Views
5K
  • · Replies 25 ·
Replies
25
Views
3K