When one hears "Monte Carlo", most people might think of something like this:
Monte Carlo, Monaco: known for extremely large amounts of money, car racing, no income taxes,and copious gambling.
In addition to Monaco, Europe, Las Vegas decided to host a Monte Carlo-themed casino as well. So during the Manhattan project, when the best minds in the United States were camped out in the New Mexican desert, they had plenty of inspiration from Las Vegas, and plenty of difficult problems to work on in the form of quantifying the inner workings of nuclei. Enrico Fermi first played with these ideas, but Stanislaw Ulam invented the modern Monte Carlo Markov Chain later.
At the same time, these scientists now had computers at their disposal. John von Neumann programmed Ulam's algorithm onto ENIAC (Electronic Numerical Integrator and Computer), the very first electronic, general purpose computer, even though it did still run on vacuum tubes.
That still doesn't answer, why do random numbers actually help us solve problems?
Imagine you are visiting a new city for the first time (maybe Monte Carlo). You only have a day or two, and you want to really get to know the city. You have two options for your visit
Both are legitimate ways to see the city. But depending on what you want, you might choose a different option. The same goes for exploring physics problems. Sometimes you want to go in and study just everything you knew about beforehand, but sometimes you need to wander around, not hit everything, but get a better feeling for what everything might be like.
Even back in the 18th century, Georges-Louis Leclerc, Comte de Buffon posed a problem in geometric probability. Nowadays, we use a slightly different version of that problem to calculate π and illustrate Monte Carlo simulations.
Suppose we have a square dartboard, and someone with really bad, completely random aim, even though he/she always at least hits inside the dartboard. We then inscribe a circle inside that dartboard. After an infinite number of hits, what is the ratio between hits in the circle, and hits in the square?
In [1]:
using Statistics
using Plots
gr()
Out[1]:
We will generate our random numbers on the unit interval. Thus the radius in our circumstance is $.5$.
Write a function incircle(r2)
such that if r2
is in the circle, it returns true, else, it returns false. We will use this with the Julia function filter
. Assume r2
is the radius squared, and already centered around the middle of the unit circle
In [2]:
function incircle(r2)
if r2<.25
return true
else
return false
end
end
Out[2]:
In [3]:
#The number of darts we will throw at the board. We will see how accurate different numbers are
N=[10,25,50,75,100,250,500,750,1000,2500,5000,7500,10000];
# We will perform each number multiple times in order to calculate error bars
M=15;
In [4]:
πapprox=zeros(Float64,length(N),M);
for ii in 1:length(N)
for jj in 1:M
#popular our array with random numbers on the unit interval
X=rand(N[ii],2)
#calculate their radius squared
R2=(X[:,1].-0.5).^2.0.+(X[:,2].-0.5).^2
# 4*number in circle / total number
πapprox[ii,jj]=4.0*length(filter(incircle,R2))/N[ii];
end
end
In [5]:
# Get our averages and standard deviations
πave=mean(πapprox,dims=2);
πstd=std(πapprox,dims=2);
In [6]:
plot(N,π*ones(length(N)),xscale=:log10);
for j in 1:M
scatter!(N,πapprox[:,j]);
end
scatter!(N,πave,yerr=πstd)
plot!(xlabel="Number of Darts",ylabel="pi Estimate",
title="Monte Carlo Estimate of pi",legend=:false)
Out[6]:
When we have fewer numbers of points, our estimates vary much more wildly, and much further from 3.1415926 . But, at least, the guesses from our different runs all seem equally distributed around the correct value, so it seems we have no systematic error.
As we get up to $10^4$, our estimate starts getting much more accurate and consistent.
In [7]:
plot(N,πstd,xscale=:log10)
plot!(xlabel="N points"
,ylabel="standard deviation"
,title="Dependence of Monte Carlo Error on Number of Points")
Out[7]:
So what we guessed in the first plot about dispersion in estimate, we quantify here in this plot. When we only have 10 darts, the guesses vary by up to .3, but when we get down to 1,000 trials, we are starting to be consistent to .0002
In [8]:
plot(N,πave,xscale=:log10,label="π Guess")
plot!(N,π*ones(length(N)),label="π")
plot!(xlabel="N steps"
,ylabel="Average of 15 runs"
,title="Overall Averages")
Out[8]:
Now lets just make a graphical representation of what we've been doing this whole time. Plot our points on a unit square, and color the ones inside the circle a different color.
In [9]:
X=rand(1000);
Y=rand(1000);
R2=(X.-0.5).^2.0.+(Y.-0.5).^2;
Xc=[];
Yc=[]
for ii in 1:length(X)
if R2[ii]<.25
push!(Xc,X[ii]);
push!(Yc,Y[ii]);
end
end
In [10]:
scatter(X,Y)
scatter!(Xc,Yc)
plot!(aspect_ratio=1,xlabel="X",ylabel="Y",legend=:false,
title="Dartboard")
Out[10]:
That's all folks! Now here's a picture of some pie to congratulate you on calculating π.
By Scott Bauer, USDA ARS - This image was released by the Agricultural Research Service, the research agency of the United States Department of Agriculture, with the ID K7252-47 (next).This tag does not indicate the copyright status of the attached work. A normal copyright tag is still required. See Commons:Licensing for more information.English | français | македонски | +/−, Public Domain, https://commons.wikimedia.org/w/index.php?curid=264106
@article{Markov, author = {A. Markov , A}, year = {2006}, month = {12}, pages = {591 - 600}, title = {An Example of Statistical Investigation of the Text Eugene Onegin Concerning the Connection of Samples in Chains}, volume = {19}, journal = {Science in Context}, doi = {10.1017/S0269889706001074} }
In [ ]: