Investigating Bell’s Theorem on Correlated Systems Utilizing Monte Carlo Methods
Document Type
Oral Presentation
Campus where you would like to present
SURC 140
Start Date
16-5-2013
End Date
16-5-2013
Abstract
Bell’s inequality is an expression which can be employed in order to test for the existence of hidden variables in a correlated system. The idea of hidden variables was introduced as a potential explanation for the quantum mechanical phenomenon of non-deterministic results of measurements. The existence of hidden variables can be tested experimentally. First, a particular measurement of one component of a correlated system must be associated with a positive or negative integer value. There must be an equal probability density of obtaining either measurement. Following this, an ensemble average of these measurements must be performed and Bell’s inequality must subsequently be employed. This theorem was essential in settling the EPR argument against quantum mechanical indeterminism, experimentally showing through violation of the inequality that there are indeed no hidden variables in a quantum system. For a system exhibiting classical behavior, Bell’s inequality should not be violated, provided the correlated system meets criteria appropriate to the assumptions underlying Bell’s theorem. A Mathematica model of Bell’s inequality was developed utilizing a Monte Carlo approach of randomizing parameters in order to test models of correlated systems. The model appears to agree with classical determinism within one standard deviation. It is possible to test both classical and quantum mechanical systems if properly correlated, and it is of future interest to test whether this extends to chaotic systems.
Recommended Citation
Corbin, Ryan, "Investigating Bell’s Theorem on Correlated Systems Utilizing Monte Carlo Methods" (2013). Symposium Of University Research and Creative Expression (SOURCE). 20.
https://digitalcommons.cwu.edu/source/2013/oralpresentations/20
Additional Mentoring Department
Physics
Investigating Bell’s Theorem on Correlated Systems Utilizing Monte Carlo Methods
SURC 140
Bell’s inequality is an expression which can be employed in order to test for the existence of hidden variables in a correlated system. The idea of hidden variables was introduced as a potential explanation for the quantum mechanical phenomenon of non-deterministic results of measurements. The existence of hidden variables can be tested experimentally. First, a particular measurement of one component of a correlated system must be associated with a positive or negative integer value. There must be an equal probability density of obtaining either measurement. Following this, an ensemble average of these measurements must be performed and Bell’s inequality must subsequently be employed. This theorem was essential in settling the EPR argument against quantum mechanical indeterminism, experimentally showing through violation of the inequality that there are indeed no hidden variables in a quantum system. For a system exhibiting classical behavior, Bell’s inequality should not be violated, provided the correlated system meets criteria appropriate to the assumptions underlying Bell’s theorem. A Mathematica model of Bell’s inequality was developed utilizing a Monte Carlo approach of randomizing parameters in order to test models of correlated systems. The model appears to agree with classical determinism within one standard deviation. It is possible to test both classical and quantum mechanical systems if properly correlated, and it is of future interest to test whether this extends to chaotic systems.
Faculty Mentor(s)
Michael Braunstein